Skip to content

[Question] 如何正常运行LLaDA1.5,现在出来的都是MASK #39

@WuNein

Description

@WuNein

Required prerequisites

Questions

from diffulex import Diffulex, SamplingParams
from transformers import AutoTokenizer



# Initialize the Diffulex engine
model_path = "/root/models/LLaDA-1.5"
llm = Diffulex(
    model_path,
    model_name="llada",  # or "dream", "llada", etc.
    tensor_parallel_size=1,
    data_parallel_size=1,
    gpu_memory_utilization=0.7,
    max_model_len=2048,
    decoding_strategy="d2f",  # or "d2f", "fast_dllm"
    mask_token_id=126336,  # model-specific mask token ID
enforce_eager = True,
kv_cache_layout= "unified",
use_lora = False,
    add_block_threshold=0.1,
    semi_complete_threshold=0.9,
decoding_threshold=0.95,
block_size=32,
buffer_size=4,
)

# Load tokenizer
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)

# Set sampling parameters
sampling_params = SamplingParams(
    temperature=0.0,
    max_tokens=128,
)



# Prepare prompts
prompts = [
    "Question: What is the capital of France? Answer:",
    "Question: Explain quantum computing in simple terms. Answer:",
]

# Generate responses
outputs = llm.generate(prompts, sampling_params)
#print(outputs.convert_to_text(tokenizer))
print(outputs.to_benchmark_format())
# Process results
#for output in outputs:
#    print(f"Generated text: {output['text']}")
#    print(f"Number of diffusion steps: {output['n_diff_steps']}")
#    print(f"Token IDs: {output['token_ids']}")

如果按照Readme来[rank0]: TypeError: 'GenerationOutputs' object is not iterable

print(outputs.to_benchmark_format()) 可以输出内容,但是

Diffulex Generating: 100%|███████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:11<00:00,  5.89s/it, tpf=1.09tok/step, ttft=1.12s, tpot=0.04s, ptps=114.47tok/s, dtps=52.07tok/s, tps=47.28tok/s]
[08:57:57] INFO     --------------------------------
           INFO     Generation Outputs Summary:
           INFO     --------------------------------
           INFO     Total Tokens: 512 toks
           INFO     Total NFEs: 254 nfes (steps)
           INFO     Total Time: 11.718297283630818 sec
           INFO     TPF: 1.09 tok/step
           INFO     TTFT: 1.12 sec
           INFO     TPOT: 0.04 sec
           INFO     Throughput: 47.28 tok/sec
           INFO     Prefill Throughput: 114.47 tok/sec
           INFO     Decode Throughput: 52.07 tok/sec
           INFO     --------------------------------
None
[{'text': '', 'full_text': '<|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|>', 'token_ids': [126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463], 'nfe': 254}, {'text': '', 'full_text': '<|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|><|mdm_mask|>', 'token_ids': [126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463, 126463], 'nfe': 254}]

结果都是MASK的空值,我已经根据benchmark yaml改了llada的配置参数,请问能否提供llada的运行配置。

Metadata

Metadata

Assignees

No one assigned

    Labels

    questionFurther information is requested

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions