🛡️video model API

reflection - StreamingT2V API Documentation


Introduction

reflection is a project built on the StreamingT2V framework, offering powerful tools for converting text input into video output in a streaming fashion. This documentation covers the API provided by the reflection project, which allows developers to convert text to video, manipulate video segments, and stream the resulting videos.

Installation

To install the reflection project, you can use the following command:

pip install reflection-streamingt2v

Basic Concepts

Before using the API, it is essential to understand some basic concepts:

  • T2VContext: Manages the global state and configuration for text-to-video operations.

  • T2VStream: Represents a streaming session for text-to-video conversion.

  • TextToVideoProcessor: Processes text input and generates video segments.

  • VideoSegment: Represents a segment of video generated from a portion of text.

  • VideoOutput: Manages the output stream of the video.

API Reference

T2VContext

class reflection.T2VContext

The T2VContext class encapsulates the global state and configuration for the reflection project.

  • Methods:

    • __init__(config: dict): Initialize a new context with the given configuration.

    • dispose(): Dispose of the context and free associated resources.

T2VStream

class reflection.T2VStream

The T2VStream class represents a streaming session for text-to-video conversion.

  • Methods:

    • __init__(context: T2VContext): Create a new streaming session within the given context.

    • start(): Start the streaming session.

    • stop(): Stop the streaming session.

    • send_text(text: str): Send text input to the streaming session.

    • receive_video_segment() -> VideoSegment: Receive a video segment generated from the text input.

TextToVideoProcessor

class reflection.TextToVideoProcessor

The TextToVideoProcessor class processes text input and generates video segments.

  • Methods:

    • __init__(context: T2VContext): Create a new text-to-video processor within the given context.

    • process_text(text: str) -> VideoSegment: Process the given text and generate a video segment.

    • set_parameter(name: str, value: any): Set a parameter for the text-to-video processor.

    • get_parameter(name: str) -> any: Get the value of a parameter for the text-to-video processor.

VideoSegment

class reflection.VideoSegment

The VideoSegment class represents a segment of video generated from a portion of text.

  • Methods:

    • __init__(data: bytes, metadata: dict): Create a new video segment with the given data and metadata.

    • get_data() -> bytes: Get the binary data of the video segment.

    • get_metadata() -> dict: Get the metadata of the video segment.

VideoOutput

class reflection.VideoOutput

The VideoOutput class manages the output stream of the video.

  • Methods:

    • __init__(output_path: str): Create a new video output to the specified path.

    • write_segment(segment: VideoSegment): Write a video segment to the output.

    • close(): Close the video output stream.

Examples

Creating a Context and Starting a Stream

Processing Text to Video

Writing Video Output

Contributing

Contributions to the reflection project are welcome. Please follow the standard GitHub workflow for contributing:

  1. Fork the repository.

  2. Create a new branch for your feature or bugfix.

  3. Commit your changes and push them to your branch.

  4. Create a pull request.

Ensure your code follows the project's coding standards and includes appropriate tests.

License

The reflection project is licensed under the MIT License. See the LICENSE file for more details.


This documentation provides an overview of the reflection API for streaming text-to-video conversion. For more detailed information and advanced usage, please refer to the source code and additional documentation in the project's repository.

Last updated