diff --git a/.NET/commands.md b/.NET/commands.md new file mode 100644 index 0000000..c04eeb5 --- /dev/null +++ b/.NET/commands.md @@ -0,0 +1,2 @@ + +Add-BindingRedirect * \ No newline at end of file diff --git a/azure/ddeployments.md b/azure/ddeployments.md new file mode 100644 index 0000000..d5f733b --- /dev/null +++ b/azure/ddeployments.md @@ -0,0 +1,2 @@ + +az resource update --resource-group default-switzerland --name ftp --namespace Microsoft.Web --resource-type basicPublishingCredentialsPolicies --parent sites/gateway-junior-test --set properties.allow=true \ No newline at end of file diff --git a/dev/prisma.md b/dev/prisma.md new file mode 100644 index 0000000..dddb228 --- /dev/null +++ b/dev/prisma.md @@ -0,0 +1,5 @@ +migrate: +# creates js classes +npx prisma generate + +# deploy \ No newline at end of file diff --git a/portainer-compose-stacks/AI/n8n/dev1.yml b/portainer-compose-stacks/AI/n8n/dev1.yml new file mode 100644 index 0000000..f6779e9 --- /dev/null +++ b/portainer-compose-stacks/AI/n8n/dev1.yml @@ -0,0 +1,131 @@ +version: '3.8' + +services: + # llava-controller: + # deploy: + # replicas: 0 + # image: db-llava:latest + # ports: + # - 10001:10000 + # command: python -m llava.serve.controller --host 0.0.0.0 --port 10000 + + # gradio-web-server: + # deploy: + # replicas: 0 + # image: db-llava:latest + # depends_on: + # - llava-controller + # ports: + # - 7861:7860 + # command: python -m llava.serve.gradio_web_server --controller http://llava-controller:10000 --model-list-mode reload + + # model-worker: + # image: db-llava:latest + # depends_on: + # - llava-controller + # ports: + # - 40000:40000 + # command: python -m llava.serve.model_worker --host 0.0.0.0 --controller http://llava-controller:10000 --port 40000 --worker http://localhost:40000 --model-path liuhaotian/llava-v1.6-vicuna-13b #liuhaotian/llava-v1.5-13b + # environment: + # NVIDIA_VISIBLE_DEVICES: all + # deploy: + # replicas: 0 + # resources: + # reservations: + # devices: + # - driver: nvidia + # count: 1 + # capabilities: [gpu] + + # h2o-llmstudio: + # image: gcr.io/vorvan/h2oai/h2o-llmstudio:nightly + # runtime: nvidia + # shm_size: 64g + # init: true + # container_name: h2o-llmstudio + # # user: "${UID}:${GID}" + # ports: + # - "10101:10101" + # volumes: + # - /mnt/storage/docker_slow/lm-studio/data:/workspace/data + # - /mnt/storage/docker_slow/lm-studio/output:/workspace/output + # - /mnt/storage/docker_slow/lm-studio/.cache:/home/llmstudio/.cache + # environment: + # - NVIDIA_VISIBLE_DEVICES=all + # - NVIDIA_DRIVER_CAPABILITIES=compute,utility + + + # # # # + # https://docs.n8n.io/hosting/installation/docker/#prerequisites +# docker volume create n8n_data + +# docker run -it --rm --name n8n -p 5678:5678 -v n8n_data:/home/node/.n8n docker.n8n.io/n8nio/n8n + + # # # + n8n: + image: betterweb/n8n-docker + ports: + - "5679:5678" + environment: + - N8N_BASIC_AUTH_ACTIVE=true + - N8N_BASIC_AUTH_USER=admin + - N8N_BASIC_AUTH_PASSWORD=9UXnZ49GR67r6VH + - N8N_HOST=${N8N_HOST:-localhost} + - N8N_PORT=5678 + - N8N_PROTOCOL=http + - N8N_ENCRYPTION_KEY=whtqTVF5C8367$74V + - DB_TYPE=postgresdb + - DB_POSTGRESDB_HOST=db + - DB_POSTGRESDB_DATABASE=n8n + - DB_POSTGRESDB_USER=n8n + - DB_POSTGRESDB_PASSWORD=n8n + volumes: + - n8n_data:/home/node/.n8n + depends_on: + - db + restart: unless-stopped + + db: + image: postgres:13 + environment: + - POSTGRES_DB=n8n + - POSTGRES_USER=n8n + - POSTGRES_PASSWORD=n8n + volumes: + - db_data:/var/lib/postgresql/data + restart: unless-stopped + +volumes: + n8n_data: + db_data: + # # # + # neo4j: + # image: neo4j + # ports: + # - "7474:7474" + # - "7687:7687" + # environment: + # NEO4J_AUTH: neo4j/lucas-bicycle-powder-stretch-ford-9492 + # NEO4JLABS_PLUGINS: '["apoc"]' + # NEO4J_apoc_export_file_enabled: 'true' + # NEO4J_apoc_import_file_enabled: 'true' + # NEO4J_apoc_import_file_use__neo4j__config: 'true' + # NEO4J_dbms_security_procedures_unrestricted: apoc.* + # volumes: + # - "/mnt/storage/docker_slow/neo4j/data:/data" + # lmdeploy: + # runtime: nvidia + # #environment: + # # - HUGGING_FACE_HUB_TOKEN=${HUGGING_FACE_HUB_TOKEN} + # volumes: + # - /mnt/storage/docker_slow/.cache/huggingface:/root/.cache/huggingface + # ports: + # - "23333:23333" + # ipc: host + # deploy: + # resources: + # reservations: + # devices: + # - capabilities: [gpu] + # command: lmdeploy serve api_server internlm/internlm2_5-7b-chat + # image: openmmlab/lmdeploy:latest \ No newline at end of file