forked from camenduru/flux.1-dev-lora-tost
-
Notifications
You must be signed in to change notification settings - Fork 0
/
runpod.toml
72 lines (63 loc) · 3.07 KB
/
runpod.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
# RunPod Project Configuration
name = "flux1-dev-lora-worker-project"
[project]
# uuid - Unique identifier for the project. Automatically generated.
#
# volume_mount_path - Default volume mount path in serverless environment. Changing this may affect data persistence.
#
# base_image - Base Docker image for the project environment. Includes essential packages and CUDA support.
# - Use 'runpod/base' as a starting point. Customize only if you need additional packages or configurations.
#
# gpu_types - List of GPU types for your development pod. Order the types from most preferred to least preferred.
# - The pod uses the first available type from this list.
# - For a full list of supported GPU types, visit: https://docs.runpod.io/references/gpu-types
#
# gpu_count - Number of GPUs to allocate for this pod.
#
# ports - Ports to expose and their protocols. Configure as needed for your application.
#
# container_disk_size_gb - Disk space allocated to the container. Adjust according to your needs.
uuid = "0fa3d603"
base_image = "timpietruskyblibla/flux-1-dev-lora-worker"
gpu_types = [
"NVIDIA GeForce RTX 4090", # 24GB
"NVIDIA A100 80GB PCIe", # 80GB
]
gpu_count = 1
volume_mount_path = "/runpod-volume"
ports = "4040/http, 7270/http, 22/tcp" # FileBrowser, FastAPI, SSH
container_disk_size_gb = 100
[project.env_vars]
# Set environment variables for the pod.
# For a full list of base environment variables, visit: https://github.com/runpod/containers/blob/main/official-templates/base/Dockerfile
#
# POD_INACTIVITY_TIMEOUT - Number of seconds before terminating the pod after a session ends. End a session with 'CTRL+C'.
# - You only pay for the pod until it terminates.
#
# RUNPOD_DEBUG_LEVEL - Log level for RunPod. Set to 'debug' for detailed logs.
#
# UVICORN_LOG_LEVEL - Log level for Uvicorn. Set to 'warning' for minimal logs.
POD_INACTIVITY_TIMEOUT = "120"
RUNPOD_DEBUG_LEVEL = "debug"
UVICORN_LOG_LEVEL = "warning"
[endpoint]
# Configure the deployed endpoint.
# For a full list of endpoint configurations, visit: https://docs.runpod.io/serverless/references/endpoint-configurations
#
# active_workers - The minimum number of workers your endpoint has running at any given point.
# - Setting this amount to 1 or more runs that number of "always on" workers.
# - These workers respond to job requests without any cold start delay.
#
# max_workers - The maximum number of workers your endpoint has running at any given point.
active_workers = 0
max_workers = 3
flashboot = true
[runtime]
# python_version - Python version to use for the project.
#
# handler_path - Path to the handler file for the project. Adapt example scripts from Hugging Face in this file.
#
# requirements_path - Path to the requirements file for the project. Add dependencies from Hugging Face in this file.
python_version = "3.10"
handler_path = "src/handler.py"
requirements_path = "builder/requirements.txt"