@@ -3,7 +3,6 @@ import { DockerService } from '#services/docker_service'
33import { BaseSeeder } from '@adonisjs/lucid/seeders'
44import { ModelAttributes } from '@adonisjs/lucid/types/model'
55import env from '#start/env'
6- import { RagService } from '#services/rag_service'
76
87export default class ServiceSeeder extends BaseSeeder {
98 // Use environment variable with fallback to production default
@@ -64,11 +63,11 @@ export default class ServiceSeeder extends BaseSeeder {
6463 } ,
6564 {
6665 service_name : DockerService . OLLAMA_SERVICE_NAME ,
67- friendly_name : 'Ollama ' ,
68- powered_by : null ,
69- display_order : 100 , // Dependency service, not shown directly
70- description : 'Run local LLMs ( AI models) with ease on your own hardware' ,
71- icon : 'IconRobot ' ,
66+ friendly_name : 'AI Assistant ' ,
67+ powered_by : 'Ollama' ,
68+ display_order : 3 ,
69+ description : 'Local AI chat that runs entirely on your hardware - no internet required ' ,
70+ icon : 'IconWand ' ,
7271 container_image : 'ollama/ollama:latest' ,
7372 container_command : 'serve' ,
7473 container_config : JSON . stringify ( {
@@ -82,40 +81,8 @@ export default class ServiceSeeder extends BaseSeeder {
8281 ui_location : null ,
8382 installed : false ,
8483 installation_status : 'idle' ,
85- is_dependency_service : true ,
86- depends_on : DockerService . QDRANT_SERVICE_NAME ,
87- } ,
88- {
89- service_name : DockerService . OPEN_WEBUI_SERVICE_NAME ,
90- friendly_name : 'AI Assistant' ,
91- powered_by : 'Open WebUI + Ollama' ,
92- display_order : 3 ,
93- description : 'Local AI chat that runs entirely on your hardware - no internet required' ,
94- icon : 'IconWand' ,
95- container_image : 'ghcr.io/open-webui/open-webui:main' ,
96- container_command : null ,
97- container_config : JSON . stringify ( {
98- HostConfig : {
99- RestartPolicy : { Name : 'unless-stopped' } ,
100- NetworkMode : 'host' ,
101- Binds : [ `${ ServiceSeeder . NOMAD_STORAGE_ABS_PATH } /open-webui:/app/backend/data` ] ,
102- PortBindings : { '8080/tcp' : [ { HostPort : '3000' } ] } ,
103- } ,
104- Env : [
105- 'WEBUI_AUTH=False' ,
106- 'PORT=3000' ,
107- 'OLLAMA_BASE_URL=http://127.0.0.1:11434' ,
108- 'VECTOR_DB=qdrant' ,
109- 'QDRANT_URI=http://127.0.0.1:6333' ,
110- 'RAG_EMBEDDING_ENGINE=ollama' ,
111- `RAG_EMBEDDING_MODEL=${ RagService . EMBEDDING_MODEL } ` ,
112- ] ,
113- } ) ,
114- ui_location : '3000' ,
115- installed : false ,
116- installation_status : 'idle' ,
11784 is_dependency_service : false ,
118- depends_on : DockerService . OLLAMA_SERVICE_NAME ,
85+ depends_on : DockerService . QDRANT_SERVICE_NAME ,
11986 } ,
12087 {
12188 service_name : DockerService . CYBERCHEF_SERVICE_NAME ,
0 commit comments