TextGenerationBenchmark Class — pytorch Architecture
Architecture documentation for the TextGenerationBenchmark class in huggingface_llm_models.py from the pytorch codebase.
Entity Profile
Relationship Graph
Source Code
benchmarks/dynamo/huggingface_llm_models.py lines 66–93
class TextGenerationBenchmark(Benchmark):
INPUT_LENGTH = 1000
OUTPUT_LENGTH = 2000
@staticmethod
def get_model_and_inputs(model_name, device):
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, device_map=device)
model.eval()
model.generation_config.do_sample = False
model.generation_config.use_cache = True
model.generation_config.cache_implementation = "static"
model.generation_config.max_new_tokens = TextGenerationBenchmark.OUTPUT_LENGTH
model.generation_config.pad_token_id = tokenizer.eos_token_id
model.generation_config.temperature = 0.0
vocab_size = tokenizer.vocab_size
input_ids = torch.randint(
low=0,
high=vocab_size,
size=(1, TextGenerationBenchmark.INPUT_LENGTH),
device=device,
dtype=torch.long,
)
example_inputs = {"input_ids": input_ids}
return model, example_inputs
Domain
Source
Analyze Your Own Codebase
Get architecture documentation, dependency graphs, and domain analysis for your codebase in minutes.
Try Supermodel Free