You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I would like to request the addition of a parameter to control the input length in the run_pipeline.py script for image-to-text examples. This would allow users to specify how the input sequences are padded or truncated, making the script more flexible for different use cases.
If --max_input_tokens > 0: Pad and truncate input sequences to the specified number of tokens.
If --max_input_tokens == 0 (default): default behavior, Do not truncate input sequences and use the full input prompt.
The primary motivation for this feature request is to enable benchmarking of the run_pipeline.py script for various input and output token lengths. Currently, the script uses a fixed length for default input, limiting its flexibility for benchmarking use cases. By introducing a parameter to control the input length, users can measure the script's performance with different input lengths to better understand how token length impacts runtime, memory usage, and overall throughput.
The text was updated successfully, but these errors were encountered:
@sywangyi, Could you please take a look at this issue. Let me know if you need any clarification or additional details. Your support on this is much appreciated!
I would like to request the addition of a parameter to control the input length in the run_pipeline.py script for image-to-text examples. This would allow users to specify how the input sequences are padded or truncated, making the script more flexible for different use cases.
If --max_input_tokens > 0: Pad and truncate input sequences to the specified number of tokens.
If --max_input_tokens == 0 (default): default behavior, Do not truncate input sequences and use the full input prompt.
The primary motivation for this feature request is to enable benchmarking of the run_pipeline.py script for various input and output token lengths. Currently, the script uses a fixed length for default input, limiting its flexibility for benchmarking use cases. By introducing a parameter to control the input length, users can measure the script's performance with different input lengths to better understand how token length impacts runtime, memory usage, and overall throughput.
The text was updated successfully, but these errors were encountered: