UNPKG

@huggingface/tasks

Version:
78 lines (77 loc) 2.06 kB
/** * Inference code generated from the JSON schema spec in ./spec * * Using src/scripts/inference-codegen */ /** * Inputs for Image Text To Video inference. Either inputs (image) or prompt (in parameters) * must be provided, or both. */ export interface ImageTextToVideoInput { /** * The input image data as a base64-encoded string. If no `parameters` are provided, you can * also provide the image data as a raw bytes payload. Either this or prompt must be * provided. */ inputs?: Blob; /** * Additional inference parameters for Image Text To Video */ parameters?: ImageTextToVideoParameters; [property: string]: unknown; } /** * Additional inference parameters for Image Text To Video */ export interface ImageTextToVideoParameters { /** * For diffusion models. A higher guidance scale value encourages the model to generate * videos closely linked to the text prompt at the expense of lower image quality. */ guidance_scale?: number; /** * One prompt to guide what NOT to include in video generation. */ negative_prompt?: string; /** * The num_frames parameter determines how many video frames are generated. */ num_frames?: number; /** * The number of denoising steps. More denoising steps usually lead to a higher quality * video at the expense of slower inference. */ num_inference_steps?: number; /** * The text prompt to guide the video generation. Either this or inputs (image) must be * provided. */ prompt?: string; /** * Seed for the random number generator. */ seed?: number; /** * The size in pixel of the output video frames. */ target_size?: TargetSize; [property: string]: unknown; } /** * The size in pixel of the output video frames. */ export interface TargetSize { height: number; width: number; [property: string]: unknown; } /** * Outputs of inference for the Image Text To Video task */ export interface ImageTextToVideoOutput { /** * The generated video returned as raw bytes in the payload. */ video: unknown; [property: string]: unknown; }