-
Notifications
You must be signed in to change notification settings - Fork 460
Expand file tree
/
Copy pathmodel_configs.yaml
More file actions
26 lines (25 loc) · 938 Bytes
/
model_configs.yaml
File metadata and controls
26 lines (25 loc) · 938 Bytes
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
device: cuda # cuda / cpu / mps (using `transformers` as backend)
weights:
PP-DocLayoutV2: Structure/PP-DocLayoutV2
layoutreader: Relation
models_dir: model_weight
layout_config:
model: PP-DocLayoutV2
reader:
name: layoutreader
chat_config:
weight_path: model_weight/Recognition
backend: lmdeploy # lmdeploy / vllm / transformers / api / lmdeploy_queue / vllm_queue / vllm_async
data_parallelism: 1 # for lmdeploy only (test)
model_parallelism: 1 # for lmdeploy and vllm
batch_size: 10 # active when using `transformers` as backend
# if using xxx_queue as backend
queue_config:
max_batch_size: 256 # maximum batch size for internal processing
queue_timeout: 1 # seconds to wait for batching requests
max_queue_size: 2000 # maximum requests in queue
# Uncomment the following lines if use `api` as backend
# api_config:
# url: https://api.openai.com/v1
# model_name: gpt-4.1
# api_key: sk-xxx