Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
24 commits
Select commit Hold shift + click to select a range
dfe03fb
update samples to use "production" environment by default
jordanstephens Jan 23, 2026
aea5750
manual down workflow
muddysound Jan 26, 2026
0d522d1
set SSL_MODE=require for crewai
jordanstephens Jan 27, 2026
01b0deb
print deployment summary in new step
jordanstephens Jan 27, 2026
7d1834b
crewai set config-vars-init-random
jordanstephens Jan 27, 2026
0c08dd3
Update samples/crewai/compose.yaml
jordanstephens Jan 27, 2026
6529ef7
Update samples/crewai/compose.yaml
jordanstephens Jan 27, 2026
fab5bdc
support down from deploy workflow
jordanstephens Jan 28, 2026
0a152a3
remove 1-click button from readme
jordanstephens Jan 28, 2026
c591b71
generate random secrets automatically
jordanstephens Jan 28, 2026
6059a2e
rename llm service to chat
jordanstephens Jan 28, 2026
2d6570f
Remove DEBUG variable from compose.yaml
lionello Jan 28, 2026
52e7b6f
Update Defang GitHub Action to version 1.4.0
lionello Jan 28, 2026
31a8c3a
Add default action 'up' for defang job
lionello Jan 28, 2026
561042b
rename LLM_ to CHAT_
lionello Jan 28, 2026
e3ed64f
rename llm service name locally
jordanstephens Jan 28, 2026
ba5941e
rename html title
jordanstephens Jan 28, 2026
5766c0c
Reduce health check intervals and timeouts
lionello Jan 28, 2026
bc89525
Workflow updates
jordanstephens Feb 2, 2026
50d5abb
restore SSL_MODE config in crewai sample
jordanstephens Feb 2, 2026
c19d2b6
prettier format all compose files
jordanstephens Feb 2, 2026
f01b605
add concurrency to defang job
jordanstephens Feb 2, 2026
51d5434
Merge remote-tracking branch 'origin/main' into jordan/update-workflows
lionello Feb 3, 2026
bfe34e7
Merge remote-tracking branch 'origin/main' into jordan/update-workflows
lionello Feb 3, 2026
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
14 changes: 7 additions & 7 deletions samples/agentic-autogen/compose.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,21 +5,21 @@ services:
context: ./src
dockerfile: Dockerfile
ports:
- 3000:3000
- 3000:3000
environment:
- MISTRAL_API_KEY=${MISTRAL_API_KEY}
- MISTRAL_API_KEY=${MISTRAL_API_KEY}
restart: unless-stopped
healthcheck:
test:
- CMD
- curl
- -f
- http://localhost:3000/health
- CMD
- curl
- -f
- http://localhost:3000/health
interval: 30s
timeout: 10s
retries: 3
deploy:
resources:
reservations:
cpus: '0.5'
cpus: "0.5"
memory: 512M
20 changes: 10 additions & 10 deletions samples/agentic-langgraph/compose.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,30 +5,30 @@ services:
context: ./app
dockerfile: Dockerfile
ports:
- target: 3000
published: 3000
mode: ingress
- target: 3000
published: 3000
mode: ingress
environment:
TAVILY_API_KEY: null
LLM_URL: http://llm/api/v1/
LLM_MODEL: default
OPENAI_API_KEY: FAKE_TOKEN
depends_on:
- llm
- llm
deploy:
resources:
reservations:
cpus: '0.5'
cpus: "0.5"
memory: 512M
llm:
environment:
- OPENAI_API_KEY=FAKE_TOKEN
- OPENAI_API_KEY=FAKE_TOKEN
image: defangio/openai-access-gateway:latest
ports:
- target: 80
published: 80
protocol: tcp
mode: host
- target: 80
published: 80
protocol: tcp
mode: host
x-defang-llm: true
deploy:
resources:
Expand Down
28 changes: 14 additions & 14 deletions samples/agentic-strands/compose.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,40 +4,40 @@ services:
build:
context: ./app
ports:
- target: 5001
published: 5001
mode: ingress
- target: 5001
published: 5001
mode: ingress
environment:
LLM_URL: http://llm/api/v1/
LLM_MODEL: default
OPENAI_API_KEY: SAMPLE_TOKEN
healthcheck:
test:
- CMD
- python3
- -c
- import urllib.request; exit(0) if urllib.request.urlopen('http://localhost:5001/health').status
== 200 else exit(1)
- CMD
- python3
- -c
- import urllib.request; exit(0) if urllib.request.urlopen('http://localhost:5001/health').status
== 200 else exit(1)
interval: 30s
timeout: 5s
retries: 3
start_period: 10s
depends_on:
- llm
- llm
deploy:
resources:
reservations:
cpus: '0.5'
cpus: "0.5"
memory: 512M
llm:
environment:
OPENAI_API_KEY: SAMPLE_TOKEN
image: defangio/openai-access-gateway
ports:
- target: 80
published: 80
protocol: tcp
mode: host
- target: 80
published: 80
protocol: tcp
mode: host
x-defang-llm: true
deploy:
resources:
Expand Down
16 changes: 8 additions & 8 deletions samples/angular-express/compose.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,25 +4,25 @@ services:
build:
context: ./frontend
ports:
- target: 80
published: 80
mode: ingress
- target: 80
published: 80
mode: ingress
restart: unless-stopped
deploy:
resources:
reservations:
memory: 64M
depends_on:
- backend
- backend
backend:
build:
context: ./backend
ports:
- target: 3000
published: 3000
mode: ingress
- target: 3000
published: 3000
mode: ingress
environment:
- PORT=3000
- PORT=3000
restart: unless-stopped
deploy:
resources:
Expand Down
18 changes: 9 additions & 9 deletions samples/arduino-wifi/compose.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,16 +8,16 @@ services:
deploy:
resources:
reservations:
cpus: '1.0'
cpus: "1.0"
memory: 512M
ports:
- mode: ingress
target: 8081
published: 8081
- mode: ingress
target: 8081
published: 8081
healthcheck:
test:
- CMD
- python3
- -c
- import sys, urllib.request; urllib.request.urlopen(sys.argv[1]).read()
- http://localhost:8081/
- CMD
- python3
- -c
- import sys, urllib.request; urllib.request.urlopen(sys.argv[1]).read()
- http://localhost:8081/
60 changes: 30 additions & 30 deletions samples/bullmq-bullboard-redis/compose.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,76 +6,76 @@ services:
context: ./board
dockerfile: Dockerfile
ports:
- mode: ingress
target: 3000
published: 3000
- mode: ingress
target: 3000
published: 3000
environment:
- REDIS=redis://redisx:6379
- BOARD_PASSWORD
- QUEUE
- REDIS=redis://redisx:6379
- BOARD_PASSWORD
- QUEUE
healthcheck:
test:
- CMD
- curl
- -f
- 127.0.0.1:3000
- CMD
- curl
- -f
- 127.0.0.1:3000
depends_on:
- redisx
- redisx
deploy:
resources:
reservations:
cpus: '0.5'
cpus: "0.5"
memory: 512M
api:
restart: unless-stopped
build:
context: ./api
dockerfile: Dockerfile
ports:
- mode: ingress
target: 3001
published: 3001
- mode: ingress
target: 3001
published: 3001
environment:
- REDIS=redis://redisx:6379
- QUEUE
- REDIS=redis://redisx:6379
- QUEUE
healthcheck:
test:
- CMD
- curl
- -f
- 127.0.0.1:3001
- CMD
- curl
- -f
- 127.0.0.1:3001
depends_on:
- redisx
- redisx
deploy:
resources:
reservations:
cpus: '0.5'
cpus: "0.5"
memory: 512M
worker:
restart: unless-stopped
deploy:
replicas: 1
resources:
reservations:
cpus: '0.5'
cpus: "0.5"
memory: 512M
build:
context: ./worker
dockerfile: Dockerfile
environment:
- REDIS=redis://redisx:6379
- QUEUE
- REDIS=redis://redisx:6379
- QUEUE
depends_on:
- redisx
- redisx
redisx:
image: redis:6.2
x-defang-redis: true
restart: unless-stopped
ports:
- mode: host
target: 6379
- mode: host
target: 6379
deploy:
resources:
reservations:
cpus: '0.5'
cpus: "0.5"
memory: 512M
Loading
Loading