éèŠãªãã€ã³ã
- macOS : ollama.com ãã .dmg ãããŠã³ããŒããããã`brew install ollama` ãå®è¡ - ãã®åŸ `ollama run llama3.2` ã§ãã£ããéå§ã
- Windows : ollama.com/download ããã€ã³ã¹ããŒã©ãŒãããŠã³ããŒããOllama ã¯ã·ã¹ãã ãã¬ã€ã§ããã¯ã°ã©ãŠã³ã ãµãŒãã¹ãšããŠå®è¡ã
- Linux : curl ã³ãã³ã 1 ã€ã§ãã¹ãŠãã€ã³ã¹ããŒã« - `curl -fsSL https://ollama.com/install.sh | sh`
- æå°èŠä»¶ : 3B ã¢ãã«ã¯ 4 GB RAMã7B ã¢ãã«ã¯ 8 GB RAMãéå§æã« GPU ã¯äžèŠã
- Ollama 㯠`http://localhost:11434` ã§ OpenAI äºæã® REST API ãå ¬é - ã³ãŒã倿Žãªãã§ä»»æã® OpenAI SDK ã¢ããªã䜿çšã§ããŸãã
ã€ã³ã¹ããŒã«åã«ïŒããŒã«ã«LLMã¯ããªãã®ãŠãŒã¹ã±ãŒã¹ã«é©ããŠããŸããïŒ
Ollamaã®ã€ã³ã¹ããŒã«ã¯5åããããŸãããGPUæ€åºã®åé¡ããã©ã€ããŒã®äžäžèŽããŸãã¯RAMå¶çŽã«ééãããšãæåã®ã¢ãã«ãæ£ããå®è¡ããã®ã«20ïœ40åãããå¯èœæ§ããããŸãã
ããŒã«ã«æšè«ãããªãã«ãšã£ãŠæ£ããéžæãã©ããäžç¢ºããªå Žåã¯ã**ããŒã«ã«ãšã¯ã©ãŠãã®å®å šãªæš©è¡¡åæãæ¯èŒããŠãã ãã** â ã¯ã©ãŠãAPIã§éå§ããæ¹ãè³¢ããããããŸããïŒ5åã§æºåå®äºããã©ãã«ã·ã¥ãŒãã£ã³ã°äžèŠïŒãå€ãã®ãŠãŒã¶ãŒã¯ã€ã³ã¹ããŒã«åŸã«ãããçºèŠããŸãïŒä»æ±ºããã®ãè¯ãã§ãããã
ããŒã«ã«ã«ã³ããããããŠãŒã¶ãŒã¯äžèšãã芧ãã ãããã¯ã©ãŠããæåã«è©äŸ¡ããŠãããŠãŒã¶ãŒã¯ãå®å šãªæ¯èŒãåç §ããŠãã ããã
Ollama ãšã¯äœãããããŠãªã䜿çšããã®ãïŒ
äžæã§è¡šããšãOllama ã¯ãªãŒãã³ãœãŒã¹èšèªã¢ãã«ïŒMistral ã Llama 2 ãªã©ïŒãã³ã³ãã¥ãŒã¿ã«ããŠã³ããŒãããŠããŒã«ã«ã§å®è¡ããããŒã«ã§ã1 ã€ã®ã³ãã³ãã§å®è¡ã§ããŸãã
Ollama ã¯å€§èŠæš¡èšèªã¢ãã«ãããŒã«ã«ã§å®è¡ãããªãŒãã³ãœãŒã¹ã®æšè«ãšã³ãžã³ã§ãã ã¢ãã«ç®¡çãllama.cpp æšè«ããã¯ãšã³ããOpenAI äºæ REST API ãåäžã®è»œéã¢ããªã±ãŒã·ã§ã³ã«çµåããŸããPython ãªããconda ç°å¢ãªããCUDA ã»ããã¢ãããªãã
Ollama 㯠Meta Llama 3.1ãMicrosoft Phi-3ãGoogle Gemma 2ãMistralãQwen2.5ãããã³ 100 以äžã®ä»ã®ã¢ãã«çšã«ã¯ã³ã³ãã³ã ããŠã³ããŒãæ©èœãåãããã¥ã¬ãŒã·ã§ã³ãããã¢ãã« ã©ã€ãã©ãªïŒollama.com/libraryïŒãä¿æããŠããŸããã¢ãã«ã¯ 1 åããŠã³ããŒããããŠãã£ã¹ã¯ã«ãã£ãã·ã¥ããã以éã®å®è¡ã¯ 5 ç§ä»¥å ã«éå§ããŸãã
Ollama ã®ä»£ããã«ã€ããŠã¯ãLocal LLM ã¯ã³ã¯ãªã㯠ã€ã³ã¹ããŒã©ãŒ ãåç §ããŠãã ãããOllama ãš LM Studio ãæ¯èŒããã«ã¯ãLM Studio ã®ã€ã³ã¹ããŒã«æ¹æ³ ãåç §ããŠãã ããã
Ollama ã macOS ã«ã€ã³ã¹ããŒã«ããã«ã¯ã©ãããã°ããã§ããïŒ
äžæã§è¡šããšãOllama 㯠macOS ã§ããã¯ã°ã©ãŠã³ã ãµãŒãã¹ãšããŠå®è¡ãããã€ã³ã¹ããŒã«ã»èµ·ååŸãããŒã«ã« API ã `http://localhost:11434` ã§ã¢ãã« ãªã¯ãšã¹ãããªãã¹ã³ã
2 ã€ã®æ¹æ³ããããŸããã€ã³ã¹ããŒã©ãŒ ããŠã³ããŒãã®æ¹ãéãã§ããHomebrew 㯠brew ã§ãœãããŠã§ã¢ã管çããå Žåã«é©ããŠããŸãã
- 1ollama.com/download ã«ã¢ã¯ã»ã¹ãããmacOS çšããŠã³ããŒãããã¯ãªãã¯ã
- 2ããŠã³ããŒããã Ollama.dmg ãã¡ã€ã«ãéããOllama ãã¢ããªã±ãŒã·ã§ã³ ãã©ã«ãã«ãã©ãã°ã
- 3ã¢ããªã±ãŒã·ã§ã³ãã Ollama ãèµ·åãLlama ã¢ã€ã³ã³ãã¡ãã¥ãŒ ããŒã«è¡šç€º - Ollama ãããã¯ã°ã©ãŠã³ã ãµãŒãã¹ãšããŠå®è¡äžã
- 4ã¿ãŒããã«ãéããŠæåã®ã¢ãã«ãå®è¡: `ollama run llama3.2`
- 5ã¢ãã«ãããŠã³ããŒãïŒllama3.2:3b çšã«çŽ 2 GBïŒããããã£ãã ããã³ããã衚瀺ãã¡ãã»ãŒãžãå ¥åã㊠Enter ããŒãæŒããŸãã
Ollama ã Homebrew ã§ macOS ã«ã€ã³ã¹ããŒã«ããã«ã¯ã©ãããã°ããã§ããïŒ
brew install ollama
# Ollama ãµãŒãã¹ãèµ·å
ollama serve &
# ã¢ãã«ãããŠã³ããŒãããŠå®è¡
ollama run llama3.2Ollama ã Windows ã«ã€ã³ã¹ããŒã«ããã«ã¯ã©ãããã°ããã§ããïŒ
- 1ollama.com/download ã«ã¢ã¯ã»ã¹ãããWindows çšããŠã³ããŒãããã¯ãªãã¯ã
- 2ããŠã³ããŒããã OllamaSetup.exe ã€ã³ã¹ããŒã©ãŒãå®è¡ãOllama 㯠%LOCALAPPDATA%\Programs\Ollama ã«ã€ã³ã¹ããŒã«ã
- 3Ollama ãèªåçã«èµ·åããã·ã¹ãã ãã¬ã€ ã¢ã€ã³ã³ãšããŠè¡šç€ºã
- 4PowerShell ãŸãã¯ã³ãã³ã ããã³ãããéããŠå®è¡: `ollama run llama3.2`
- 5ã¢ãã«ã¯ååå®è¡æã«ããŠã³ããŒãã以éã®å®è¡ã¯ãã£ãã·ã¥ãããã¢ãã«ã䜿çšã
Windows ã§ GPU ãµããŒããæå¹ã«ããã«ã¯ã©ãããã°ããã§ããïŒ
Windows äžã® Ollama 㯠NVIDIA GPUïŒCUDA 11.3+ïŒãš AMD GPUïŒROCm 6+ïŒãèªåçã«æ€åºããŠäœ¿çšãNVIDIA RTX ã«ãŒããããå ŽåãOllama ã¯ã¢ãã« ã¬ã€ã€ãŒã VRAM ã«èªåçã«ãªãããŒã - æåèšå®ã¯äžèŠãGPU ã䜿çšãããŠããããšã確èªããã«ã¯ã`ollama run llama3.2` ãå®è¡ããŠã¿ã¹ã¯ ãããŒãžã£ãŒ â GPU ã§ç¢ºèªã
Ollama ã Linux ã«ã€ã³ã¹ããŒã«ããã«ã¯ã©ãããã°ããã§ããïŒ
1 ã€ã®ã³ãã³ãã§ä»»æã® Linux ãã£ã¹ããªãã¥ãŒã·ã§ã³ã« Ollama ãã€ã³ã¹ããŒã«:
curl -fsSL https://ollama.com/install.sh | shLinux ã§ Ollama ã systemd ãµãŒãã¹ãšããŠå®è¡ããã«ã¯ã©ãããã°ããã§ããïŒ
ã€ã³ã¹ããŒã« ã¹ã¯ãªããã¯èªåçã« Ollama ã systemd ãµãŒãã¹ãšããŠç»é²ãç®¡çæ¹æ³ã¯:
# ãµãŒãã¹ ã¹ããŒã¿ã¹ã確èª
systemctl status ollama
# éå§ / 忢 / åèµ·å
systemctl start ollama
systemctl stop ollama
systemctl restart ollama
# ãã°ã衚瀺
journalctl -u ollama -fOllama ã§æåã®ã¢ãã«ãããŠã³ããŒãããŠå®è¡ããã«ã¯ã©ãããã°ããã§ããïŒ
Ollama ã®ã€ã³ã¹ããŒã«åŸããã®ã³ãã³ããå®è¡ããŠã¢ãã«ãããŠã³ããŒãããŠèµ·å:
# ã¢ãã«ãããŠã³ããŒãïŒ~/.ollama/models ã«ä¿åïŒ
ollama pull llama3.2
# 察話çã«å®è¡
ollama run llama3.2
# ãŸã㯠1 ã¹ãããã§ããŠã³ããŒãããŠå®è¡
ollama run llama3.2æåã«äœ¿çšãã¹ãã¢ãã«ã¯ã©ããïŒ
ååå®è¡ã§ã¯ããããã® 3 ã€ã®ã¢ãã«ã¯ããŸããŸãªããŒããŠã§ã¢ ãããã¡ã€ã«ã«å¯Ÿå¿:
| ã¢ãã« | ããŠã³ããŒã ãµã€ãº | å¿ èŠãª RAM | çšé |
|---|---|---|---|
| Llama 3.2 3B | ~2 GB | 4 GB | ååãã¹ã - ãã¹ãŠã®ãã·ã³ |
| Llama 3.1 8B | ~4.7 GB | 8 GB | ã»ãšãã©ã®ããŒãããœã³ã³ã§äžè¬çã«äœ¿çš |
| phi4-mini | ~2.3 GB | 4 GB | é«éå¿çãäœ RAM |
Ollama ãåäœããŠããããšã確èªããã«ã¯ã©ãããã°ããã§ããïŒ
REST API ãçŽæ¥ãã¹ãããŠãOllama ãå®è¡ãããã¢ã¯ã»ã¹å¯èœã§ããããšã確èª:
# Ollama ãå®è¡äžã§ããããšã確èª
curl http://localhost:11434
# æåŸ
: "Ollama is running"
# ããŠã³ããŒããããã¢ãã«ããªã¹ã
ollama list
# API çµç±ã§ããã³ãããéä¿¡ïŒOpenAI äºæïŒ
curl http://localhost:11434/api/generate -d '{
"model": "llama3.2",
"prompt": "2+2 ã¯äœã§ããïŒ",
"stream": false
}'æã䟿å©ãª Ollama ã³ãã³ãã¯ã©ãã§ããïŒ
| ã³ãã³ã | æ©èœ |
|---|---|
| ollama list | ããŠã³ããŒãããããã¹ãŠã®ã¢ãã«ãšãã®ãµã€ãºã衚瀺 |
| ollama pull <model> | ã¢ãã«ãå®è¡ããã«ããŠã³ããŒã |
| ollama rm <model> | ãã£ã¹ã¯ããã¢ãã«ãåé€ |
| ollama ps | çŸåšã¡ã¢ãªã«èªã¿èŸŒãŸããŠããã¢ãã«ã衚瀺 |
| ollama show <model> | ã¢ãã«ã®è©³çްã衚瀺ïŒãã©ã¡ãŒã¿ããã³ãã¬ãŒããã©ã€ã»ã³ã¹ïŒ |
| ollama serve | Ollama ãµãŒããŒãæåã§éå§ïŒãµãŒãã¹ãšããŠå®è¡ãããŠããªãå ŽåïŒ |
Ollama ã€ã³ã¹ããŒã«æã®äžè¬çãªåé¡ããã©ãã«ã·ã¥ãŒãã£ã³ã°ããã«ã¯ã©ãããã°ããã§ããïŒ
Ollama ããcould not connect to ollama app, is it running?ãïŒOllama ã¢ããªã«æ¥ç¶ã§ããŸãããå®è¡ããŠããŸããïŒïŒãšè¡šç€º
Ollama ãããã¯ã°ã©ãŠã³ã ãµãŒãã¹ãšããŠå®è¡ãããŠããŸãããmacOS ã§ã¢ããªã±ãŒã·ã§ã³ãã Ollama ã¢ããªãéããLinux ã§ãã¿ãŒããã«ã§ `systemctl start ollama` ãŸã㯠`ollama serve` ãå®è¡ãWindows ã§ã¹ã¿ãŒã ã¡ãã¥ãŒãã Ollama ãèµ·åã
ã¢ãã«ã®ããŠã³ããŒããéåžžã«é ãããŸãã¯ã¹ã¿ãã¯ç¶æ
ã¢ãã« ããŠã³ããŒãã¯å€§ããïŒ2ïœ47 GBïŒãããŠã³ããŒãã忢ããå ŽåãCtrl+C ãæŒã㊠`ollama pull <model>` ãåå®è¡ - Ollama ã¯éšåããŠã³ããŒããåéãããéãããŠã³ããŒãã«ã¯ Wi-Fi ã§ã¯ãªãæç·æ¥ç¶ã䜿çšã
ãerror: model requires more system memoryãïŒãšã©ãŒïŒã¢ãã«ã«ã¯ã·ã¹ãã ã¡ã¢ãªãå¿ èŠïŒãšã©ãŒã衚瀺
ã¢ãã«ãå©çšå¯èœãª RAM ã«å¯ŸããŠå€§ããããŸããããå°ããéåã詊ã: ããã©ã«ãã® Q4_K_M ã§ã¯ãªã `ollama run llama3.2-instruct-q4_0`ããŸã㯠`llama3.2:3b` ã®ãããªå°ããã¢ãã«ã«åãæ¿ããåå¿è åãã®æé«ã®ããŒã«ã« LLM ã¢ãã« ãåç §ã㊠RAM ã«å¯Ÿå¿ããæšå¥šäºé ã
Ollama ã¯å®è¡ããŠããŸãããGPU ã䜿çšãããŠããŸãã
Windows ã§ NVIDIA ãã©ã€ããŒãããŒãžã§ã³ 452.39 以äžã§ããããšã確èªãLinux ã§ NVIDIA ã³ã³ãã ããŒã«ããããã€ã³ã¹ããŒã«ãããŠããããšã確èªïŒ`nvidia-smi` 㯠GPU æ å ±ãè¿ãå¿ èŠããããŸãïŒãOllama 㯠VRAM ãå©çšå¯èœãªå Žåãã¬ã€ã€ãŒã GPU ã«èªåçã«ãªãããŒã - ã¢ãã«ãéå§ããåŸ `ollama ps` ãå®è¡ã㊠GPU 䜿çšçã確èªã
Ollama ã¢ãã« ãã¡ã€ã«ã¯ã©ãã«ä¿åãããŠããŸããïŒ
ã¢ãã«ã¯ macOS ãš Linux äžã® ~/.ollama/models ã«ä¿åãWindows ã§ã¯æ¢å®ã®ãã¹ã¯ C:\Users\<username>\.ollama\modelsãOLLAMA_MODELS ç°å¢å€æ°ãèšå®ããŠãµãŒãã¹éå§åã«ã¹ãã¬ãŒãžã®å Žæã倿Žå¯èœã
Ollama ã€ã³ã¹ããŒã«åŸã¯äœãããå¿ èŠããããŸããïŒ
Ollama ãå®è¡ããå§ããããæ¬¡ã®ã¹ããã㯠æåã®ããŒã«ã« LLM ãå®è¡ ããŠããã³ããäœæãã³ã³ããã¹ãé·ãããã³äºæ³ãããããŒã«ã«æšè«é床ãçè§£ãããŒããŠã§ã¢ã«æé©ãªã¢ãã«ãéžæããã«ã¯ãåå¿è åãã®æé«ã®ããŒã«ã« LLM ã¢ãã« ãåç §ãã¿ãŒããã«ããã°ã©ãã£ã«ã«ãªãã£ãã ã€ã³ã¿ãŒãã§ãŒã¹ãåªå ããå ŽåãLM Studio ã®ã€ã³ã¹ããŒã«æ¹æ³ ã¯ãã¹ã¯ããã ã¢ããªã®ä»£æ¿ãã«ããŒã
å°åå¥ãããã€ã¡ã³ããšã³ã³ãã©ã€ã¢ã³ã¹èæ¯
æ¥æ¬ / METI: æ¥æ¬ METI AI ã¬ããã³ã¹ ã¬ã€ãã©ã€ã³ã§ã¯ AI æšè«ãçºçããå Žæã®ææžåãèŠæ±ãçµç¹ã¯ãç¹ã«éèãµãŒãã¹ãå»çã補é ã«ãããããŒã¿æåšå°ã«ãŒã«éµå®ã®ãã Ollama ããªã³ãµã€ãã§å±éå¯èœãOllama ããŒã«ã«å±é㯠APPIïŒå人æ å ±ä¿è·æ¹éæ³ïŒèŠä»¶ã«åèŽãããã©ã€ãã·ãŒä¿å šã¢ãããŒããšããŠèªèã
EU / GDPR: Ollama ããªã³ãµã€ã ã§å®è¡ããããšã¯å人ããŒã¿ãçµç¹ã®ã€ã³ãã©ã¹ãã©ã¯ãã£ãé¢ããªãããšãæå³ãGDPR èšäº 5ïŒããŒã¿æå°åïŒã¯æšè«ããã·ã³äžã§çºçããããæ¢å®ã§æºããããŸãããã€ãããã³ãã©ã³ã¹ã®ããŒã¿ä¿è·æ©é¢ã¯åŸæ¥å¡ãŸãã¯ã«ã¹ã¿ã㌠ããŒã¿åŠçã®ããããŒã«ã« LLM ãããã€ã¡ã³ããæšå¥šãOllama ã `localhost` ã«ãã€ã³ãïŒæ¢å®å€ïŒã¯å€éšãããã¯ãŒã¯ ã¢ã¯ã»ã¹ãä¿èšŒããŸããã
äžåœ / CAC: äžåœã®ãµã€ããŒç©ºé管çå±ã®çæ AI ãµãŒãã¹æ«å®æªçœ®ïŒ2023ïŒã¯äžåœãŠãŒã¶ãŒã«æäŸããã AI ãµãŒãã¹ãèŠå¶ããªã³ãµã€ã ã§å®å šã«å®è¡ããã Ollama ããŒã«ã« ãããã€ã¡ã³ã㯠CAC ãããã€ããŒå®çŸ©å€ã§ãããæ©å¯ããŒã¿ãåŠçãããšã³ã¿ãŒãã©ã€ãºäœ¿çšã±ãŒã¹ã®ã³ã³ãã©ã€ã¢ã³ã¹è² æ ãå€§å¹ ã«äœæžã
ãœãŒã¹
- Ollama å ¬åŒããã¥ã¡ã³ã - ã€ã³ã¹ããŒã« ã¬ã€ããš API ããã¥ã¡ã³ã
- Ollama GitHub ãªããžã㪠- ãœãŒã¹ ã³ãŒããåé¡ãããã³ãªãªãŒã¹ ããŒã
- Ollama ã¢ãã« ã©ã€ãã©ãª - å©çšå¯èœãªã¢ãã«ã®å®å šãªãªã¹ããšããŠã³ããŒã ãªã³ã¯
Ollama ã€ã³ã¹ããŒã«æã®äžè¬çãªééããšã¯ïŒ
- Ollama ãããã¯ã°ã©ãŠã³ã ãµãŒãã¹ãšããŠå®è¡ãããŠããããšã確èªããã«ãAPI ãå¿çãããšäºæ³ãmacOS ã§ã¯ ã¡ãã¥ãŒ ããŒã« Llama ã¢ã€ã³ã³ã衚瀺ãããããšã確èªãLinux ã§ `systemctl status ollama` ãå®è¡ãWindows ã§ã·ã¹ãã ãã¬ã€ã確èªã
- æåã«ã¡ã¢ãªèŠä»¶ã確èªããã«ãå©çšå¯èœãª RAM ãã倧ããã¢ãã«ãå®è¡ããããšããã7B ã¢ãã«ã Q4 éåã«ã¯çŽ 4ïœ5 GB VRAM ãå¿ èŠãã¢ãã« ãã¡ã€ã« ãµã€ãºã« 1.2 ãåžžã«ä¹ããŠå¿ èŠãª RAM ãæšå®ã
- GPU æ€åºãç¡èŠ - Ollama 㯠NVIDIA ãš AMD ããµããŒãäœãææ°ã®ãã©ã€ããŒãå¿ èŠãWindows ã§ NVIDIA ãã©ã€ã㌠ããŒãžã§ã³ 452.39+ ã `nvidia-smi` ã§ç¢ºèªãLinux ã§ NVIDIA ã³ã³ãã ããŒã«ããããã€ã³ã¹ããŒã«ãããŠããããšã確èªã
- VRAMïŒGPU ã¡ã¢ãªïŒãšã·ã¹ãã RAM ãæ··åãOllama 㯠CPU ã§å®è¡ã§ãäœã GPU å éã 5ïœ10 åé«éã颿£ GPU ããããæšè«ãé ãå ŽåãOllama 㯠GPU ãã©ã€ããŒãäžè¶³ãŸãã¯æä»£é ãã®ãã CPU ã«ãã©ãŒã« ããã¯ã
- ã¢ãã« ããŠã³ããŒãããã£ãã·ã¥ãããŠããããšãçè§£ããŠããªããååå®è¡æ `ollama pull llama3.2` 㯠2 GB ããŠã³ããŒãã« 5ïœ10 åã以éã®å®è¡ã¯ãã£ãã·ã¥ãããã¢ãã«ã䜿çšã㊠5 ç§ä»¥å ã«éå§ã
ãããã質å
Ollama ã¯ç¡æã§ããïŒ
ã¯ããOllama 㯠MIT ã©ã€ã»ã³ã¹äžã§ç¡æãã€ãªãŒãã³ãœãŒã¹ã䜿çšå¶éãAPI ããŒäžèŠããã¹ãŠã®æšè«ã¯ãã·ã³ã§ããŒã«ã«ã«å®è¡ã
Ollama 㯠Windows ã§åäœããŸããïŒ
ã¯ããOllama 㯠2024 幎以éãå®å®ãããã€ãã£ã Windows ã€ã³ã¹ããŒã©ãŒãæã€ (2026 幎ã«ããã©ãŒãã³ã¹æ¹å)ãollama.com ããããŠã³ããŒããWindows ã§ NVIDIA GPU ã CUDA çµç±ã§ãAMD GPU ã ROCm çµç±ã§ãµããŒãã
Ollama ãå®è¡ããã®ã«å¿ èŠãª RAM ã¯ã©ããããã§ããïŒ
Q4 éåã® 3Bïœ7B ã¢ãã«ã«æå° 8 GB RAMã16 GB RAM 㯠7B ã¢ãã«ãå¿«é©ã«åŠçã13B ã¢ãã«ã Q4 ã§ãCPU ã®ã¿ã® 34B ã¢ãã«ã«ã¯ 32 GB+ RAM ããå§ãã
Ollama ãææ°ããŒãžã§ã³ã«æŽæ°ããã«ã¯ã©ãããã°ããã§ããïŒ
macOS ã§ã¯ Ollama ã¯èªåæŽæ°ãWindows ã§ã¯ ollama.com ããææ°ã€ã³ã¹ããŒã©ãŒãããŠã³ããŒãããŠå®è¡ãLinux ã§ã¯ã€ã³ã¹ããŒã« ã¹ã¯ãªãããåå®è¡: curl -fsSL https://ollama.com/install.sh | sh
ã³ãŒã倿Žãªãã§ OpenAI SDK çµç±ã§ Ollama ã䜿çšã§ããŸããïŒ
ã¯ããOpenAI SDK ã§ base_url ã http://localhost:11434/v1 ã«èšå®ããä»»æã®æååã API ããŒãšããŠæž¡ããOllama ã® REST API ã¯å®å šã« OpenAI äºæãªã®ã§ãGPT ãŸã㯠Claude åãã«èšè¿°ãããã¢ããªã±ãŒã·ã§ã³ã¯ããŒã«ã« ã¢ãã«ã䜿çšã§ããŸãã
Ollama æšè«ãé ãïŒ5 ããŒã¯ã³/ç§æªæºïŒã®ã¯ãªãã§ããïŒ
ã¢ãã«ã¯ GPU ã§ã¯ãªã CPU ã§å®è¡å¯èœãollama ps ã§ã¢ãã«ãèªã¿èŸŒãŸããŠããããšã確èªãGPU 䜿çšçã 0% ã®å ŽåãGPU ãã©ã€ããŒãã€ã³ã¹ããŒã«ã»ææ°åãããŠããããšã確èªãNVIDIA ã§ã¯ nvidia-smi ã GPU ã衚瀺ãã¹ããAMD ã§ã¯ rocm-smiãMac ã§ã¯ Metal GPU å éã Apple Silicon ã§èªåã
Ollama ã¯è€æ°ã®ã¢ãã«ãåæã«å®è¡ã§ããŸããïŒ
Ollama ã¯ããã»ã¹ãããäžåºŠã« 1 ã€ã®ã¢ãã«ãå®è¡ããã ããè€æ°ã® Ollama ã€ã³ã¹ã¿ã³ã¹ãç°ãªãããŒãïŒäŸïŒOLLAMA_HOST=localhost:11434 ããã³ OLLAMA_HOST=localhost:11435ïŒã§å®è¡ããè€æ°ã®ã¢ãã«ã䞊åã§æäŸå¯èœãããã«ã¯ ãããªã RAM ãå¿ èŠã
ollama pull ãš ollama run ã®éãã¯ïŒ
ollama pull 㯠Ollama ã©ã€ãã©ãªããã¢ãã«ãããŠã³ããŒããã¡ã¢ãªã«èªã¿èŸŒãŸããollama run ã¯ã¢ãã«ãããŠã³ããŒãïŒãã£ãã·ã¥ãããŠããªãå ŽåïŒããå³åº§ã«ãã£ãã ã»ãã·ã§ã³éå§ãAPI çµç±ã§ãã£ãã ã€ã³ã¿ãŒãã§ãŒã¹ãªãã«ã¢ãã«ã䜿çšããã«ã¯ãæåã« pull ããŠãã API ãã¯ãšãªã
Ollama ã䜿çšããããŒã«ã« LLM ã¯å人æ å ±ä¿è·æ¹éæ³ïŒAPPIïŒæºæ ã§ããïŒ
ãã¹ãŠã§ã¯ãããŸããããäžéšã§ããããŒã«ã« LLM 㯠ããŒã¿ãããŒã«ã«ã«åŠçãäœã APPI æºæ ã«ã¯ ããã«: ãã£ã¹ã¯æå·åãã¢ã¯ã»ã¹å¶åŸ¡ãåŠçèšé²ãããã³è©²åœããå Žåã¯æ¥åå§èšå¥çŽãå¿ èŠãOllama ã«ããããŒã«ã«åŠçã¯éèŠãªã¹ãããäœã APPI æºæ ã®å®å šãªèšŒæ ã§ã¯ãªããããŒã¿ä¿è·è²¬ä»»è ïŒDPOïŒã«çžè«ã
æ¥æ¬ã®å€§äŒæ¥ã§ Ollama ãæ¬çªç°å¢ã§äœ¿çšã§ããŸããïŒ
ã¯ããå€ãã®æ¥æ¬äŒæ¥ã¯å éšããã¥ã¡ã³ãåŠçãã³ãŒã ã¬ãã¥ãŒãããŒã¿åæã®ãã Ollama ãããŒã«ã«ã§äœ¿çšãéèŠ: Ollama ã localhost ã«ãã€ã³ãïŒæ¢å®å€ïŒãæ©å¯ããŒã¿çšã«ãã£ã¹ã¯æå·åãæå¹åãã³ã³ãã©ã€ã¢ã³ã¹èŠä»¶ã確èªïŒç¹ã«ã«ã¹ã¿ã㌠ããŒã¿ïŒããããã®å¯Ÿçã«ãããOllama ã¯ã¯ã©ãŠãäŸåãªãã«çµç¹å ã§ AI ãå©çšããè²»çšå¹ççãªæ¹æ³ã