Support path objs in components (#4581)

* Added support for Path objs in components(Video, Geallery, chatbot)

* added tests for pathlib.Path objs in Video, chatbot, Gallery

* added changes to docstrings and type annotations

* changelog

* formatting

* video

* fixing video docstring

* update

* Update test/test_components.py

* Update test/test_components.py

* formatting

* remaining components

* add pathlib to tests

* fix test

* formatting

---------

Co-authored-by: Abubakar Abid <abubakar@huggingface.co>
This commit is contained in:
Sunil Kumar Dash 2023-06-30 03:24:35 +05:30 committed by GitHub
parent 9361a29da2
commit e38496ec1e
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
8 changed files with 87 additions and 26 deletions

View File

@ -5,6 +5,7 @@
- Allow the web component `space`, `src`, and `host` attributes to be updated dynamically by [@pngwn](https://github.com/pngwn) in [PR 4461](https://github.com/gradio-app/gradio/pull/4461)
- Spaces Duplication built into Gradio, by [@aliabid94](https://github.com/aliabid94) in [PR 4458](https://github.com/gradio-app/gradio/pull/4458)
- The `api_name` parameter now accepts `False` as a value, which means it does not show up in named or unnamed endpoints. By [@abidlabs](https://github.com/aliabid94) in [PR 4683](https://github.com/gradio-app/gradio/pull/4683)
- Added support for `pathlib.Path` in `gr.Video`, `gr.Gallery`, and `gr.Chatbot` by [sunilkumardash9](https://github.com/sunilkumardash9) in [PR 4581](https://github.com/gradio-app/gradio/pull/4581).
## Bug Fixes:

View File

@ -42,7 +42,7 @@ class Audio(
"""
Creates an audio component that can be used to upload/record audio (as an input) or display audio (as an output).
Preprocessing: passes the uploaded audio as a {Tuple(int, numpy.array)} corresponding to (sample rate in Hz, audio data as a 16-bit int array whose values range from -32768 to 32767), or as a {str} filepath, depending on `type`.
Postprocessing: expects a {Tuple(int, numpy.array)} corresponding to (sample rate in Hz, audio data as a float or int numpy array) or as a {str} filepath or URL to an audio file, which gets displayed
Postprocessing: expects a {Tuple(int, numpy.array)} corresponding to (sample rate in Hz, audio data as a float or int numpy array) or as a {str} or {pathlib.Path} filepath or URL to an audio file, which gets displayed
Examples-format: a {str} filepath to a local file that contains audio.
Demos: main_note, generate_tone, reverse_audio
Guides: real-time-speech-recognition
@ -50,7 +50,7 @@ class Audio(
def __init__(
self,
value: str | tuple[int, np.ndarray] | Callable | None = None,
value: str | Path | tuple[int, np.ndarray] | Callable | None = None,
*,
source: Literal["upload", "microphone"] = "upload",
type: Literal["numpy", "filepath"] = "numpy",
@ -300,7 +300,9 @@ class Audio(
masked_inputs.append(masked_data)
return masked_inputs
def postprocess(self, y: tuple[int, np.ndarray] | str | None) -> str | dict | None:
def postprocess(
self, y: tuple[int, np.ndarray] | str | Path | None
) -> str | dict | None:
"""
Parameters:
y: audio data in either of the following formats: a tuple of (sample_rate, data), or a string filepath or URL to an audio file, or None.

View File

@ -179,7 +179,7 @@ class IOComponent(Component):
self.attach_load_event(load_fn, every)
@staticmethod
def hash_file(file_path: str, chunk_num_blocks: int = 128) -> str:
def hash_file(file_path: str | Path, chunk_num_blocks: int = 128) -> str:
sha1 = hashlib.sha1()
with open(file_path, "rb") as f:
for chunk in iter(lambda: f.read(chunk_num_blocks * sha1.block_size), b""):
@ -214,7 +214,7 @@ class IOComponent(Component):
sha1.update(data.encode("utf-8"))
return sha1.hexdigest()
def make_temp_copy_if_needed(self, file_path: str) -> str:
def make_temp_copy_if_needed(self, file_path: str | Path) -> str:
"""Returns a temporary file path for a copy of the given file path if it does
not already exist. Otherwise returns the path to the existing temp file."""
temp_dir = self.hash_file(file_path)

View File

@ -3,6 +3,7 @@
from __future__ import annotations
import inspect
from pathlib import Path
from typing import Callable, Literal
from gradio_client import utils as client_utils
@ -26,7 +27,7 @@ class Chatbot(Changeable, Selectable, IOComponent, JSONSerializable):
"""
Displays a chatbot output showing both user submitted messages and responses. Supports a subset of Markdown including bold, italics, code, and images.
Preprocessing: this component does *not* accept input.
Postprocessing: expects function to return a {List[List[str | None | Tuple]]}, a list of lists. The inner list should have 2 elements: the user message and the response message. Messages should be strings, tuples, or Nones. If the message is a string, it can include Markdown. If it is a tuple, it should consist of (string filepath to image/video/audio, [optional string alt text]). Messages that are `None` are not displayed.
Postprocessing: expects function to return a {List[List[str | None | Tuple]]}, a list of lists. The inner list should have 2 elements: the user message and the response message. Messages should be strings, tuples, or Nones. If the message is a string, it can include Markdown. If it is a tuple, it should consist of (filepath to image/video/audio, [optional string alt text]). Messages that are `None` are not displayed.
Demos: chatbot_simple, chatbot_multimodal
Guides: creating-a-chatbot
@ -34,7 +35,7 @@ class Chatbot(Changeable, Selectable, IOComponent, JSONSerializable):
def __init__(
self,
value: list[list[str | tuple[str] | tuple[str, str] | None]]
value: list[list[str | tuple[str] | tuple[str | Path, str] | None]]
| Callable
| None = None,
color_map: dict[str, str] | None = None,
@ -171,7 +172,7 @@ class Chatbot(Changeable, Selectable, IOComponent, JSONSerializable):
if chat_message is None:
return None
elif isinstance(chat_message, (tuple, list)):
file_uri = chat_message[0]
file_uri = str(chat_message[0])
if utils.validate_url(file_uri):
filepath = file_uri
else:
@ -197,7 +198,7 @@ class Chatbot(Changeable, Selectable, IOComponent, JSONSerializable):
) -> list[list[str | dict | None]]:
"""
Parameters:
y: List of lists representing the message and response pairs. Each message and response should be a string, which may be in Markdown format. It can also be a tuple whose first element is a string filepath or URL to an image/video/audio, and second (optional) element is the alt text, in which case the media file is displayed. It can also be None, in which case that message is not displayed.
y: List of lists representing the message and response pairs. Each message and response should be a string, which may be in Markdown format. It can also be a tuple whose first element is a string or pathlib.Path filepath or URL to an image/video/audio, and second (optional) element is the alt text, in which case the media file is displayed. It can also be None, in which case that message is not displayed.
Returns:
List of lists representing the message and response. Each message and response will be a string of HTML, or a dictionary with media information. Or None if the message is not to be displayed.
"""

View File

@ -2,6 +2,7 @@
from __future__ import annotations
from pathlib import Path
from typing import Any, Callable, Literal
import numpy as np
@ -25,14 +26,16 @@ class Gallery(IOComponent, GallerySerializable, Selectable):
"""
Used to display a list of images as a gallery that can be scrolled through.
Preprocessing: this component does *not* accept input.
Postprocessing: expects a list of images in any format, {List[numpy.array | PIL.Image | str]}, or a {List} of (image, {str} caption) tuples and displays them.
Postprocessing: expects a list of images in any format, {List[numpy.array | PIL.Image | str | pathlib.Path]}, or a {List} of (image, {str} caption) tuples and displays them.
Demos: fake_gan
"""
def __init__(
self,
value: list[np.ndarray | _Image.Image | str | tuple] | Callable | None = None,
value: list[np.ndarray | _Image.Image | str | Path | tuple]
| Callable
| None = None,
*,
label: str | None = None,
every: float | None = None,
@ -172,7 +175,7 @@ class Gallery(IOComponent, GallerySerializable, Selectable):
file = self.pil_to_temp_file(img, dir=self.DEFAULT_TEMP_DIR)
file_path = str(utils.abspath(file))
self.temp_files.add(file_path)
elif isinstance(img, str):
elif isinstance(img, (str, Path)):
if utils.validate_url(img):
file_path = img
else:

View File

@ -26,8 +26,8 @@ class Model3D(
):
"""
Component allows users to upload or view 3D Model files (.obj, .glb, or .gltf).
Preprocessing: This component passes the uploaded file as a {str} filepath.
Postprocessing: expects function to return a {str} path to a file of type (.obj, glb, or .gltf)
Preprocessing: This component passes the uploaded file as a {str}filepath.
Postprocessing: expects function to return a {str} or {pathlib.Path} filepath of type (.obj, glb, or .gltf)
Demos: model3D
Guides: how-to-use-3D-model-component
@ -135,7 +135,7 @@ class Model3D(
return temp_file_path
def postprocess(self, y: str | None) -> dict[str, str] | None:
def postprocess(self, y: str | Path | None) -> dict[str, str] | None:
"""
Parameters:
y: path to the model

View File

@ -41,14 +41,18 @@ class Video(
that the output video would not be playable in the browser it will attempt to convert it to a playable mp4 video.
If the conversion fails, the original video is returned.
Preprocessing: passes the uploaded video as a {str} filepath or URL whose extension can be modified by `format`.
Postprocessing: expects a {str} filepath to a video which is displayed, or a {Tuple[str, str]} where the first element is a filepath to a video and the second element is a filepath to a subtitle file.
Postprocessing: expects a {str} or {pathlib.Path} filepath to a video which is displayed, or a {Tuple[str | pathlib.Path, str | pathlib.Path | None]} where the first element is a filepath to a video and the second element is an optional filepath to a subtitle file.
Examples-format: a {str} filepath to a local file that contains the video, or a {Tuple[str, str]} where the first element is a filepath to a video file and the second element is a filepath to a subtitle file.
Demos: video_identity, video_subtitle
"""
def __init__(
self,
value: str | tuple[str, str | None] | Callable | None = None,
value: str
| Path
| tuple[str | Path, str | Path | None]
| Callable
| None = None,
*,
format: str | None = None,
source: Literal["upload", "webcam"] = "upload",
@ -234,13 +238,12 @@ class Video(
return str(file_name)
def postprocess(
self, y: str | tuple[str, str | None] | None
self, y: str | Path | tuple[str | Path, str | Path | None] | None
) -> tuple[FileData | None, FileData | None] | None:
"""
Processes a video to ensure that it is in the correct format before
returning it to the front end.
Processes a video to ensure that it is in the correct format before returning it to the front end.
Parameters:
y: video data in either of the following formats: a tuple of (str video filepath, str subtitle filepath), or a string filepath or URL to an video file, or None.
y: video data in either of the following formats: a tuple of (video filepath, optional subtitle filepath), or just a filepath or URL to an video file, or None.
Returns:
a tuple with the two dictionary, reresent to video and (optional) subtitle, which following formats:
- The first dictionary represents the video file and contains the following keys:
@ -257,12 +260,15 @@ class Video(
if y is None or y == [None, None] or y == (None, None):
return None
if isinstance(y, str):
if isinstance(y, (str, Path)):
processed_files = (self._format_video(y), None)
elif isinstance(y, (tuple, list)):
assert (
len(y) == 2
), f"Expected lists of length 2 or tuples of length 2. Received: {y}"
assert isinstance(y[0], (str, Path)) and isinstance(
y[1], (str, Path)
), f"If a tuple is provided, both elements must be strings or Path objects. Received: {y}"
video = y[0]
subtitle = y[1]
processed_files = (
@ -274,7 +280,7 @@ class Video(
return processed_files
def _format_video(self, video: str | None) -> FileData | None:
def _format_video(self, video: str | Path | None) -> FileData | None:
"""
Processes a video to ensure that it is in the correct format.
Parameters:
@ -288,6 +294,7 @@ class Video(
"""
if video is None:
return None
video = str(video)
returned_format = video.split(".")[-1].lower()
if self.format is None or returned_format == self.format:
conversion_needed = False

View File

@ -893,7 +893,7 @@ class TestAudio:
).endswith(".wav")
output1 = audio_output.postprocess(y_audio.name)
output2 = audio_output.postprocess(y_audio.name)
output2 = audio_output.postprocess(Path(y_audio.name))
assert output1 == output2
def test_serialize(self):
@ -1375,6 +1375,43 @@ class TestVideo:
)
).endswith(".mp4")
p_video = gr.Video()
video_with_subtitle = gr.Video()
postprocessed_video = p_video.postprocess(Path(y_vid_path))
postprocessed_video_with_subtitle = video_with_subtitle.postprocess(
(Path(y_vid_path), Path(subtitles_path))
)
processed_video = (
{
"name": "video_sample.mp4",
"data": None,
"is_file": True,
"orig_name": "video_sample.mp4",
},
None,
)
processed_video_with_subtitle = (
{
"name": "video_sample.mp4",
"data": None,
"is_file": True,
"orig_name": "video_sample.mp4",
},
{"name": None, "data": True, "is_file": False},
)
postprocessed_video[0]["name"] = os.path.basename(
postprocessed_video[0]["name"]
)
assert processed_video == postprocessed_video
postprocessed_video_with_subtitle[0]["name"] = os.path.basename(
postprocessed_video_with_subtitle[0]["name"]
)
if postprocessed_video_with_subtitle[1]["data"]:
postprocessed_video_with_subtitle[1]["data"] = True
assert processed_video_with_subtitle == postprocessed_video_with_subtitle
def test_in_interface(self):
"""
Interface, process
@ -1862,6 +1899,9 @@ class TestChatbot:
[("test/test_files/video_sample.mp4",), "cool video"],
[("test/test_files/audio_sample.wav",), "cool audio"],
[("test/test_files/bus.png", "A bus"), "cool pic"],
[(Path("test/test_files/video_sample.mp4"),), "cool video"],
[(Path("test/test_files/audio_sample.wav"),), "cool audio"],
[(Path("test/test_files/bus.png"), "A bus"), "cool pic"],
]
processed_multimodal_msg = [
[
@ -1894,7 +1934,7 @@ class TestChatbot:
},
"cool pic",
],
]
] * 2
postprocessed_multimodal_msg = chatbot.postprocess(multimodal_msg)
postprocessed_multimodal_msg_base_names = []
for x, y in postprocessed_multimodal_msg:
@ -2068,7 +2108,7 @@ class TestModel3D:
file = "test/test_files/Box.gltf"
output1 = component.postprocess(file)
output2 = component.postprocess(file)
output2 = component.postprocess(Path(file))
assert output1 == output2
def test_in_interface(self):
@ -2167,6 +2207,13 @@ class TestGallery:
data_restored = [d[0]["data"] for d in data_restored]
assert sorted(data) == sorted(data_restored)
postprocessed_gallery = gallery.postprocess([Path("test/test_files/bus.png")])
processed_gallery = [{"name": "bus.png", "data": None, "is_file": True}]
postprocessed_gallery[0]["name"] = os.path.basename(
postprocessed_gallery[0]["name"]
)
assert processed_gallery == postprocessed_gallery
class TestState:
def test_as_component(self):