mirror of
https://github.com/nextcloud/all-in-one.git
synced 2026-02-04 04:56:52 +00:00
Compare commits
2 commits
5d30f0f82d
...
df10e43da4
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
df10e43da4 | ||
|
|
808ff2f76e |
5 changed files with 27 additions and 99 deletions
|
|
@ -14,7 +14,7 @@ This container bundles caddy and auto-configures it for you. It also covers [vau
|
||||||
- If you want to use this with [nocodb](https://github.com/nextcloud/all-in-one/tree/main/community-containers/nocodb), make sure that you point `tables.your-nc-domain.com` to your server using a cname record so that caddy can get a certificate automatically for nocodb.
|
- If you want to use this with [nocodb](https://github.com/nextcloud/all-in-one/tree/main/community-containers/nocodb), make sure that you point `tables.your-nc-domain.com` to your server using a cname record so that caddy can get a certificate automatically for nocodb.
|
||||||
- If you want to use this with [jellyseerr](https://github.com/nextcloud/all-in-one/tree/main/community-containers/jellyseerr), make sure that you point `requests.your-nc-domain.com` to your server using a cname record so that caddy can get a certificate automatically for jellyseerr.
|
- If you want to use this with [jellyseerr](https://github.com/nextcloud/all-in-one/tree/main/community-containers/jellyseerr), make sure that you point `requests.your-nc-domain.com` to your server using a cname record so that caddy can get a certificate automatically for jellyseerr.
|
||||||
- If you want to use this with [nextcloud-exporter](https://github.com/nextcloud/all-in-one/tree/main/community-containers/nextcloud-exporter), make sure that you point `metrics.your-nc-domain.com` to your server using a cname record so that caddy can get a certificate automatically for nextcloud-exporter.
|
- If you want to use this with [nextcloud-exporter](https://github.com/nextcloud/all-in-one/tree/main/community-containers/nextcloud-exporter), make sure that you point `metrics.your-nc-domain.com` to your server using a cname record so that caddy can get a certificate automatically for nextcloud-exporter.
|
||||||
- If you want to use this with [local AI with vulkan support](https://github.com/nextcloud/all-in-one/tree/main/community-containers/local-ai-vulkan), make sure that you point `ai.your-nc-domain.com` to your server using a cname record so that caddy can get a certificate automatically for local AI.
|
- If you want to use this with [local AI](https://github.com/nextcloud/all-in-one/tree/main/community-containers/local-ai), make sure that you point `ai.your-nc-domain.com` to your server using a cname record so that caddy can get a certificate automatically for local AI.
|
||||||
- After the container was started the first time, you should see a new `nextcloud-aio-caddy` folder and inside there an `allowed-countries.txt` file when you open the files app with the default `admin` user. In there you can adjust the allowed country codes for caddy by adding them to the first line, e.g. `IT FR` would allow access from italy and france. Private ip-ranges are always allowed. Additionally, in order to activate this config, you need to get an account at https://dev.maxmind.com/geoip/geolite2-free-geolocation-data and download the `GeoLite2-Country.mmdb` and upload it with this exact name into the `nextcloud-aio-caddy` folder. Afterwards restart all containers from the AIO interface and your new config should be active!
|
- After the container was started the first time, you should see a new `nextcloud-aio-caddy` folder and inside there an `allowed-countries.txt` file when you open the files app with the default `admin` user. In there you can adjust the allowed country codes for caddy by adding them to the first line, e.g. `IT FR` would allow access from italy and france. Private ip-ranges are always allowed. Additionally, in order to activate this config, you need to get an account at https://dev.maxmind.com/geoip/geolite2-free-geolocation-data and download the `GeoLite2-Country.mmdb` and upload it with this exact name into the `nextcloud-aio-caddy` folder. Afterwards restart all containers from the AIO interface and your new config should be active!
|
||||||
- You can add your own Caddy configurations in `/data/caddy-imports/` inside the Caddy container (`sudo docker exec -it nextcloud-aio-caddy bash`). These will be imported on container startup. **Please note:** If you do not have CLI access to the server, you can now run docker commands via a web session by using this community container: https://github.com/nextcloud/all-in-one/tree/main/community-containers/container-management
|
- You can add your own Caddy configurations in `/data/caddy-imports/` inside the Caddy container (`sudo docker exec -it nextcloud-aio-caddy bash`). These will be imported on container startup. **Please note:** If you do not have CLI access to the server, you can now run docker commands via a web session by using this community container: https://github.com/nextcloud/all-in-one/tree/main/community-containers/container-management
|
||||||
- See https://github.com/nextcloud/all-in-one/tree/main/community-containers#community-containers how to add it to the AIO stack
|
- See https://github.com/nextcloud/all-in-one/tree/main/community-containers#community-containers how to add it to the AIO stack
|
||||||
|
|
|
||||||
|
|
@ -1,50 +0,0 @@
|
||||||
{
|
|
||||||
"aio_services_v1": [
|
|
||||||
{
|
|
||||||
"container_name": "nextcloud-aio-local-ai-vulkan",
|
|
||||||
"display_name": "Local AI for Vulkan (x86 only)",
|
|
||||||
"documentation": "https://github.com/nextcloud/all-in-one/tree/main/community-containers/local-ai-vulkan",
|
|
||||||
"image": "ghcr.io/docjyj/aio-local-ai-vulkan",
|
|
||||||
"image_tag": "v0",
|
|
||||||
"internal_port": "8080",
|
|
||||||
"restart": "unless-stopped",
|
|
||||||
"environment": [
|
|
||||||
"TZ=%TIMEZONE%",
|
|
||||||
"LOCAL_AI_PASS=%LOCAL_AI_PASS%"
|
|
||||||
],
|
|
||||||
"ports": [
|
|
||||||
{
|
|
||||||
"ip_binding": "%APACHE_IP_BINDING%",
|
|
||||||
"port_number": "10078",
|
|
||||||
"protocol": "tcp"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"volumes": [
|
|
||||||
{
|
|
||||||
"source": "nextcloud_aio_localai_vulkan_models",
|
|
||||||
"destination": "/models",
|
|
||||||
"writeable": true
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"source": "nextcloud_aio_localai_vulkan_backends",
|
|
||||||
"destination": "/backends",
|
|
||||||
"writeable": true
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"secrets": [
|
|
||||||
"LOCAL_AI_PASS"
|
|
||||||
],
|
|
||||||
"ui_secret": "LOCAL_AI_PASS",
|
|
||||||
"devices": [
|
|
||||||
"/dev/dri"
|
|
||||||
],
|
|
||||||
"nextcloud_exec_commands": [
|
|
||||||
"php /var/www/html/occ app:install integration_openai",
|
|
||||||
"php /var/www/html/occ app:enable integration_openai",
|
|
||||||
"php /var/www/html/occ config:app:set integration_openai url --value http://nextcloud-aio-local-ai-vulkan:8080",
|
|
||||||
"php /var/www/html/occ app:install assistant",
|
|
||||||
"php /var/www/html/occ app:enable assistant"
|
|
||||||
]
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
|
|
@ -1,20 +0,0 @@
|
||||||
|
|
||||||
> [!WARNING]
|
|
||||||
> ARM devices are not supported by this container. It is only for x86_64 devices. (See: https://github.com/mudler/LocalAI/issues/5778)
|
|
||||||
|
|
||||||
## Local AI with Vulkan support
|
|
||||||
This container bundles Local AI and auto-configures it for you. It support hardware acceleration with Vulkan.
|
|
||||||
|
|
||||||
### Notes
|
|
||||||
Documentation is available on the container repository. This documentation is regularly updated and is intended to be as simple and detailed as possible. Thanks for all your feedback!
|
|
||||||
|
|
||||||
- See https://github.com/docjyJ/aio-local-ai-vulkan#getting-started for getting start with this container.
|
|
||||||
- See [this guide](https://github.com/nextcloud/all-in-one/discussions/5430) for how to improve AI task pickup speed
|
|
||||||
- See https://github.com/nextcloud/all-in-one/tree/main/community-containers#community-containers how to add it to the AIO stack
|
|
||||||
- Note that Nextcloud supports only one server for AI queries, so this container cannot be used at the same time as other AI containers.
|
|
||||||
|
|
||||||
### Repository
|
|
||||||
https://github.com/docjyJ/aio-local-ai-vulkan
|
|
||||||
|
|
||||||
### Maintainer
|
|
||||||
https://github.com/docjyJ
|
|
||||||
|
|
@ -3,14 +3,21 @@
|
||||||
{
|
{
|
||||||
"container_name": "nextcloud-aio-local-ai",
|
"container_name": "nextcloud-aio-local-ai",
|
||||||
"display_name": "Local AI",
|
"display_name": "Local AI",
|
||||||
"documentation": "https://github.com/nextcloud/all-in-one/tree/main/community-containers/local-ai",
|
"documentation": "https://github.com/nextcloud/all-in-one/tree/main/community-containers/local-ai-vulkan",
|
||||||
"image": "ghcr.io/szaimen/aio-local-ai",
|
"image": "ghcr.io/docjyj/aio-local-ai-vulkan",
|
||||||
"image_tag": "v2",
|
"image_tag": "v0",
|
||||||
"internal_port": "8080",
|
"internal_port": "8080",
|
||||||
"restart": "unless-stopped",
|
"restart": "unless-stopped",
|
||||||
"environment": [
|
"environment": [
|
||||||
"TZ=%TIMEZONE%",
|
"TZ=%TIMEZONE%",
|
||||||
"MODELS_PATH=/models"
|
"LOCAL_AI_PASS=%LOCAL_AI_PASS%"
|
||||||
|
],
|
||||||
|
"ports": [
|
||||||
|
{
|
||||||
|
"ip_binding": "%APACHE_IP_BINDING%",
|
||||||
|
"port_number": "10078",
|
||||||
|
"protocol": "tcp"
|
||||||
|
}
|
||||||
],
|
],
|
||||||
"volumes": [
|
"volumes": [
|
||||||
{
|
{
|
||||||
|
|
@ -19,22 +26,19 @@
|
||||||
"writeable": true
|
"writeable": true
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"source": "nextcloud_aio_localai_images",
|
"source": "nextcloud_aio_localai_backends",
|
||||||
"destination": "/tmp/generated/images/",
|
"destination": "/backends",
|
||||||
"writeable": true
|
"writeable": true
|
||||||
},
|
|
||||||
{
|
|
||||||
"source": "%NEXTCLOUD_DATADIR%",
|
|
||||||
"destination": "/nextcloud",
|
|
||||||
"writeable": false
|
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
"enable_nvidia_gpu": false,
|
"secrets": [
|
||||||
|
"LOCAL_AI_PASS"
|
||||||
|
],
|
||||||
|
"ui_secret": "LOCAL_AI_PASS",
|
||||||
|
"devices": [
|
||||||
|
"/dev/dri"
|
||||||
|
],
|
||||||
"nextcloud_exec_commands": [
|
"nextcloud_exec_commands": [
|
||||||
"mkdir '/mnt/ncdata/admin/files/nextcloud-aio-local-ai'",
|
|
||||||
"touch '/mnt/ncdata/admin/files/nextcloud-aio-local-ai/models.yaml'",
|
|
||||||
"echo 'Scanning nextcloud-aio-local-ai folder for admin user...'",
|
|
||||||
"php /var/www/html/occ files:scan --path='/admin/files/nextcloud-aio-local-ai'",
|
|
||||||
"php /var/www/html/occ app:install integration_openai",
|
"php /var/www/html/occ app:install integration_openai",
|
||||||
"php /var/www/html/occ app:enable integration_openai",
|
"php /var/www/html/occ app:enable integration_openai",
|
||||||
"php /var/www/html/occ config:app:set integration_openai url --value http://nextcloud-aio-local-ai:8080",
|
"php /var/www/html/occ config:app:set integration_openai url --value http://nextcloud-aio-local-ai:8080",
|
||||||
|
|
|
||||||
|
|
@ -1,22 +1,16 @@
|
||||||
## Local AI
|
## Local AI
|
||||||
This container bundles Local AI and auto-configures it for you.
|
This container bundles Local AI and auto-configures it for you. It support hardware acceleration with Vulkan.
|
||||||
|
|
||||||
### Notes
|
### Notes
|
||||||
- Make sure to have enough storage space available. This container alone needs ~7GB storage. Every model that you add to `models.yaml` will of course use additional space which adds up quite fast.
|
Documentation is available on the container repository. This documentation is regularly updated and is intended to be as simple and detailed as possible. Thanks for all your feedback!
|
||||||
- After the container was started the first time, you should see a new `nextcloud-aio-local-ai` folder when you open the files app with the default `admin` user. In there you should see a `models.yaml` config file. You can now add models in there. Please refer [here](https://github.com/mudler/LocalAI/blob/master/gallery/index.yaml) where you can get further urls that you can put in there. Afterwards restart all containers from the AIO interface and the models should automatically get downloaded by the local-ai container and activated.
|
|
||||||
- Example for content of `models.yaml` (if you add all of them, it takes around 10GB additional space):
|
- See https://github.com/docjyJ/aio-local-ai-vulkan#getting-started for getting start with this container.
|
||||||
```yaml
|
|
||||||
# Stable Diffusion in NCNN with c++, supported txt2img and img2img
|
|
||||||
- url: github:mudler/LocalAI/blob/master/gallery/stablediffusion.yaml
|
|
||||||
name: Stable_diffusion
|
|
||||||
```
|
|
||||||
- To make it work, you first need to browse `https://your-nc-domain.com/settings/admin/ai` and enable or disable specific features for your models in the openAI settings. Afterwards using the Nextcloud Assistant should work.
|
|
||||||
- See [this guide](https://github.com/nextcloud/all-in-one/discussions/5430) for how to improve AI task pickup speed
|
- See [this guide](https://github.com/nextcloud/all-in-one/discussions/5430) for how to improve AI task pickup speed
|
||||||
- See https://github.com/nextcloud/all-in-one/tree/main/community-containers#community-containers how to add it to the AIO stack
|
- See https://github.com/nextcloud/all-in-one/tree/main/community-containers#community-containers how to add it to the AIO stack
|
||||||
- Note that Nextcloud supports only one server for AI queries, so this container cannot be used at the same time as other AI containers.
|
- Note that Nextcloud supports only one server for AI queries, so this container cannot be used at the same time as other AI containers.
|
||||||
|
|
||||||
### Repository
|
### Repository
|
||||||
https://github.com/szaimen/aio-local-ai
|
https://github.com/docjyJ/aio-local-ai-vulkan
|
||||||
|
|
||||||
### Maintainer
|
### Maintainer
|
||||||
https://github.com/szaimen
|
https://github.com/docjyJ
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue