mirror of
https://github.com/godotengine/godot.git
synced 2024-12-15 10:12:40 +08:00
d95794ec8a
As many open source projects have started doing it, we're removing the current year from the copyright notice, so that we don't need to bump it every year. It seems like only the first year of publication is technically relevant for copyright notices, and even that seems to be something that many companies stopped listing altogether (in a version controlled codebase, the commits are a much better source of date of publication than a hardcoded copyright statement). We also now list Godot Engine contributors first as we're collectively the current maintainers of the project, and we clarify that the "exclusive" copyright of the co-founders covers the timespan before opensourcing (their further contributions are included as part of Godot Engine contributors). Also fixed "cf." Frenchism - it's meant as "refer to / see".
707 lines
19 KiB
C++
707 lines
19 KiB
C++
/**************************************************************************/
|
|
/* video_stream_theora.cpp */
|
|
/**************************************************************************/
|
|
/* This file is part of: */
|
|
/* GODOT ENGINE */
|
|
/* https://godotengine.org */
|
|
/**************************************************************************/
|
|
/* Copyright (c) 2014-present Godot Engine contributors (see AUTHORS.md). */
|
|
/* Copyright (c) 2007-2014 Juan Linietsky, Ariel Manzur. */
|
|
/* */
|
|
/* Permission is hereby granted, free of charge, to any person obtaining */
|
|
/* a copy of this software and associated documentation files (the */
|
|
/* "Software"), to deal in the Software without restriction, including */
|
|
/* without limitation the rights to use, copy, modify, merge, publish, */
|
|
/* distribute, sublicense, and/or sell copies of the Software, and to */
|
|
/* permit persons to whom the Software is furnished to do so, subject to */
|
|
/* the following conditions: */
|
|
/* */
|
|
/* The above copyright notice and this permission notice shall be */
|
|
/* included in all copies or substantial portions of the Software. */
|
|
/* */
|
|
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
|
|
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
|
|
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. */
|
|
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
|
|
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
|
|
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
|
|
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
|
|
/**************************************************************************/
|
|
|
|
#include "video_stream_theora.h"
|
|
|
|
#include "core/config/project_settings.h"
|
|
#include "core/os/os.h"
|
|
|
|
#ifdef _MSC_VER
|
|
#pragma warning(push)
|
|
#pragma warning(disable : 4127)
|
|
#endif
|
|
|
|
#include "thirdparty/misc/yuv2rgb.h"
|
|
|
|
#ifdef _MSC_VER
|
|
#pragma warning(pop)
|
|
#endif
|
|
|
|
int VideoStreamPlaybackTheora::buffer_data() {
|
|
char *buffer = ogg_sync_buffer(&oy, 4096);
|
|
|
|
#ifdef THEORA_USE_THREAD_STREAMING
|
|
|
|
int read;
|
|
|
|
do {
|
|
thread_sem->post();
|
|
read = MIN(ring_buffer.data_left(), 4096);
|
|
if (read) {
|
|
ring_buffer.read((uint8_t *)buffer, read);
|
|
ogg_sync_wrote(&oy, read);
|
|
} else {
|
|
OS::get_singleton()->delay_usec(100);
|
|
}
|
|
|
|
} while (read == 0);
|
|
|
|
return read;
|
|
|
|
#else
|
|
|
|
uint64_t bytes = file->get_buffer((uint8_t *)buffer, 4096);
|
|
ogg_sync_wrote(&oy, bytes);
|
|
return (bytes);
|
|
|
|
#endif
|
|
}
|
|
|
|
int VideoStreamPlaybackTheora::queue_page(ogg_page *page) {
|
|
if (theora_p) {
|
|
ogg_stream_pagein(&to, page);
|
|
if (to.e_o_s) {
|
|
theora_eos = true;
|
|
}
|
|
}
|
|
if (vorbis_p) {
|
|
ogg_stream_pagein(&vo, page);
|
|
if (vo.e_o_s) {
|
|
vorbis_eos = true;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::video_write() {
|
|
th_ycbcr_buffer yuv;
|
|
th_decode_ycbcr_out(td, yuv);
|
|
|
|
int pitch = 4;
|
|
frame_data.resize(size.x * size.y * pitch);
|
|
{
|
|
uint8_t *w = frame_data.ptrw();
|
|
char *dst = (char *)w;
|
|
|
|
if (px_fmt == TH_PF_444) {
|
|
yuv444_2_rgb8888((uint8_t *)dst, (uint8_t *)yuv[0].data, (uint8_t *)yuv[1].data, (uint8_t *)yuv[2].data, size.x, size.y, yuv[0].stride, yuv[1].stride, size.x << 2);
|
|
|
|
} else if (px_fmt == TH_PF_422) {
|
|
yuv422_2_rgb8888((uint8_t *)dst, (uint8_t *)yuv[0].data, (uint8_t *)yuv[1].data, (uint8_t *)yuv[2].data, size.x, size.y, yuv[0].stride, yuv[1].stride, size.x << 2);
|
|
|
|
} else if (px_fmt == TH_PF_420) {
|
|
yuv420_2_rgb8888((uint8_t *)dst, (uint8_t *)yuv[0].data, (uint8_t *)yuv[1].data, (uint8_t *)yuv[2].data, size.x, size.y, yuv[0].stride, yuv[1].stride, size.x << 2);
|
|
}
|
|
|
|
format = Image::FORMAT_RGBA8;
|
|
}
|
|
|
|
Ref<Image> img = memnew(Image(size.x, size.y, 0, Image::FORMAT_RGBA8, frame_data)); //zero copy image creation
|
|
|
|
texture->update(img); //zero copy send to rendering server
|
|
|
|
frames_pending = 1;
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::clear() {
|
|
if (file.is_null()) {
|
|
return;
|
|
}
|
|
|
|
if (vorbis_p) {
|
|
ogg_stream_clear(&vo);
|
|
if (vorbis_p >= 3) {
|
|
vorbis_block_clear(&vb);
|
|
vorbis_dsp_clear(&vd);
|
|
}
|
|
vorbis_comment_clear(&vc);
|
|
vorbis_info_clear(&vi);
|
|
vorbis_p = 0;
|
|
}
|
|
if (theora_p) {
|
|
ogg_stream_clear(&to);
|
|
th_decode_free(td);
|
|
th_comment_clear(&tc);
|
|
th_info_clear(&ti);
|
|
theora_p = 0;
|
|
}
|
|
ogg_sync_clear(&oy);
|
|
|
|
#ifdef THEORA_USE_THREAD_STREAMING
|
|
thread_exit = true;
|
|
thread_sem->post(); //just in case
|
|
thread.wait_to_finish();
|
|
ring_buffer.clear();
|
|
#endif
|
|
|
|
theora_p = 0;
|
|
vorbis_p = 0;
|
|
videobuf_ready = 0;
|
|
frames_pending = 0;
|
|
videobuf_time = 0;
|
|
theora_eos = false;
|
|
vorbis_eos = false;
|
|
|
|
file.unref();
|
|
playing = false;
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::set_file(const String &p_file) {
|
|
ERR_FAIL_COND(playing);
|
|
ogg_packet op;
|
|
th_setup_info *ts = nullptr;
|
|
|
|
file_name = p_file;
|
|
file = FileAccess::open(p_file, FileAccess::READ);
|
|
ERR_FAIL_COND_MSG(file.is_null(), "Cannot open file '" + p_file + "'.");
|
|
|
|
#ifdef THEORA_USE_THREAD_STREAMING
|
|
thread_exit = false;
|
|
thread_eof = false;
|
|
//pre-fill buffer
|
|
int to_read = ring_buffer.space_left();
|
|
uint64_t read = file->get_buffer(read_buffer.ptr(), to_read);
|
|
ring_buffer.write(read_buffer.ptr(), read);
|
|
|
|
thread.start(_streaming_thread, this);
|
|
#endif
|
|
|
|
ogg_sync_init(&oy);
|
|
|
|
/* init supporting Vorbis structures needed in header parsing */
|
|
vorbis_info_init(&vi);
|
|
vorbis_comment_init(&vc);
|
|
|
|
/* init supporting Theora structures needed in header parsing */
|
|
th_comment_init(&tc);
|
|
th_info_init(&ti);
|
|
|
|
theora_eos = false;
|
|
vorbis_eos = false;
|
|
|
|
/* Ogg file open; parse the headers */
|
|
/* Only interested in Vorbis/Theora streams */
|
|
int stateflag = 0;
|
|
|
|
int audio_track_skip = audio_track;
|
|
|
|
while (!stateflag) {
|
|
int ret = buffer_data();
|
|
if (ret == 0) {
|
|
break;
|
|
}
|
|
while (ogg_sync_pageout(&oy, &og) > 0) {
|
|
ogg_stream_state test;
|
|
|
|
/* is this a mandated initial header? If not, stop parsing */
|
|
if (!ogg_page_bos(&og)) {
|
|
/* don't leak the page; get it into the appropriate stream */
|
|
queue_page(&og);
|
|
stateflag = 1;
|
|
break;
|
|
}
|
|
|
|
ogg_stream_init(&test, ogg_page_serialno(&og));
|
|
ogg_stream_pagein(&test, &og);
|
|
ogg_stream_packetout(&test, &op);
|
|
|
|
/* identify the codec: try theora */
|
|
if (!theora_p && th_decode_headerin(&ti, &tc, &ts, &op) >= 0) {
|
|
/* it is theora */
|
|
memcpy(&to, &test, sizeof(test));
|
|
theora_p = 1;
|
|
} else if (!vorbis_p && vorbis_synthesis_headerin(&vi, &vc, &op) >= 0) {
|
|
/* it is vorbis */
|
|
if (audio_track_skip) {
|
|
vorbis_info_clear(&vi);
|
|
vorbis_comment_clear(&vc);
|
|
ogg_stream_clear(&test);
|
|
vorbis_info_init(&vi);
|
|
vorbis_comment_init(&vc);
|
|
|
|
audio_track_skip--;
|
|
} else {
|
|
memcpy(&vo, &test, sizeof(test));
|
|
vorbis_p = 1;
|
|
}
|
|
} else {
|
|
/* whatever it is, we don't care about it */
|
|
ogg_stream_clear(&test);
|
|
}
|
|
}
|
|
/* fall through to non-bos page parsing */
|
|
}
|
|
|
|
/* we're expecting more header packets. */
|
|
while ((theora_p && theora_p < 3) || (vorbis_p && vorbis_p < 3)) {
|
|
/* look for further theora headers */
|
|
int ret = ogg_stream_packetout(&to, &op);
|
|
while (theora_p && theora_p < 3 && ret) {
|
|
if (ret < 0) {
|
|
fprintf(stderr, "Error parsing Theora stream headers; corrupt stream?\n");
|
|
clear();
|
|
return;
|
|
}
|
|
if (!th_decode_headerin(&ti, &tc, &ts, &op)) {
|
|
fprintf(stderr, "Error parsing Theora stream headers; corrupt stream?\n");
|
|
clear();
|
|
return;
|
|
}
|
|
ret = ogg_stream_packetout(&to, &op);
|
|
theora_p++;
|
|
}
|
|
|
|
/* look for more vorbis header packets */
|
|
ret = ogg_stream_packetout(&vo, &op);
|
|
while (vorbis_p && vorbis_p < 3 && ret) {
|
|
if (ret < 0) {
|
|
fprintf(stderr, "Error parsing Vorbis stream headers; corrupt stream?\n");
|
|
clear();
|
|
return;
|
|
}
|
|
ret = vorbis_synthesis_headerin(&vi, &vc, &op);
|
|
if (ret) {
|
|
fprintf(stderr, "Error parsing Vorbis stream headers; corrupt stream?\n");
|
|
clear();
|
|
return;
|
|
}
|
|
vorbis_p++;
|
|
if (vorbis_p == 3) {
|
|
break;
|
|
}
|
|
ret = ogg_stream_packetout(&vo, &op);
|
|
}
|
|
|
|
/* The header pages/packets will arrive before anything else we
|
|
care about, or the stream is not obeying spec */
|
|
|
|
if (ogg_sync_pageout(&oy, &og) > 0) {
|
|
queue_page(&og); /* demux into the appropriate stream */
|
|
} else {
|
|
int ret2 = buffer_data(); /* someone needs more data */
|
|
if (ret2 == 0) {
|
|
fprintf(stderr, "End of file while searching for codec headers.\n");
|
|
clear();
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* And now we have it all. Initialize decoders. */
|
|
if (theora_p) {
|
|
td = th_decode_alloc(&ti, ts);
|
|
px_fmt = ti.pixel_fmt;
|
|
switch (ti.pixel_fmt) {
|
|
case TH_PF_420:
|
|
//printf(" 4:2:0 video\n");
|
|
break;
|
|
case TH_PF_422:
|
|
//printf(" 4:2:2 video\n");
|
|
break;
|
|
case TH_PF_444:
|
|
//printf(" 4:4:4 video\n");
|
|
break;
|
|
case TH_PF_RSVD:
|
|
default:
|
|
printf(" video\n (UNKNOWN Chroma sampling!)\n");
|
|
break;
|
|
}
|
|
th_decode_ctl(td, TH_DECCTL_GET_PPLEVEL_MAX, &pp_level_max,
|
|
sizeof(pp_level_max));
|
|
pp_level = 0;
|
|
th_decode_ctl(td, TH_DECCTL_SET_PPLEVEL, &pp_level, sizeof(pp_level));
|
|
pp_inc = 0;
|
|
|
|
int w;
|
|
int h;
|
|
w = ((ti.pic_x + ti.frame_width + 1) & ~1) - (ti.pic_x & ~1);
|
|
h = ((ti.pic_y + ti.frame_height + 1) & ~1) - (ti.pic_y & ~1);
|
|
size.x = w;
|
|
size.y = h;
|
|
|
|
Ref<Image> img = Image::create_empty(w, h, false, Image::FORMAT_RGBA8);
|
|
texture->set_image(img);
|
|
|
|
} else {
|
|
/* tear down the partial theora setup */
|
|
th_info_clear(&ti);
|
|
th_comment_clear(&tc);
|
|
}
|
|
|
|
th_setup_free(ts);
|
|
|
|
if (vorbis_p) {
|
|
vorbis_synthesis_init(&vd, &vi);
|
|
vorbis_block_init(&vd, &vb);
|
|
//_setup(vi.channels, vi.rate);
|
|
} else {
|
|
/* tear down the partial vorbis setup */
|
|
vorbis_info_clear(&vi);
|
|
vorbis_comment_clear(&vc);
|
|
}
|
|
|
|
playing = false;
|
|
buffering = true;
|
|
time = 0;
|
|
audio_frames_wrote = 0;
|
|
}
|
|
|
|
double VideoStreamPlaybackTheora::get_time() const {
|
|
// FIXME: AudioServer output latency was fixed in af9bb0e, previously it used to
|
|
// systematically return 0. Now that it gives a proper latency, it broke this
|
|
// code where the delay compensation likely never really worked.
|
|
return time - /* AudioServer::get_singleton()->get_output_latency() - */ delay_compensation;
|
|
}
|
|
|
|
Ref<Texture2D> VideoStreamPlaybackTheora::get_texture() const {
|
|
return texture;
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::update(double p_delta) {
|
|
if (file.is_null()) {
|
|
return;
|
|
}
|
|
|
|
if (!playing || paused) {
|
|
//printf("not playing\n");
|
|
return;
|
|
}
|
|
|
|
#ifdef THEORA_USE_THREAD_STREAMING
|
|
thread_sem->post();
|
|
#endif
|
|
|
|
time += p_delta;
|
|
|
|
if (videobuf_time > get_time()) {
|
|
return; //no new frames need to be produced
|
|
}
|
|
|
|
bool frame_done = false;
|
|
bool audio_done = !vorbis_p;
|
|
|
|
while (!frame_done || (!audio_done && !vorbis_eos)) {
|
|
//a frame needs to be produced
|
|
|
|
ogg_packet op;
|
|
bool no_theora = false;
|
|
bool buffer_full = false;
|
|
|
|
while (vorbis_p && !audio_done && !buffer_full) {
|
|
int ret;
|
|
float **pcm;
|
|
|
|
/* if there's pending, decoded audio, grab it */
|
|
ret = vorbis_synthesis_pcmout(&vd, &pcm);
|
|
if (ret > 0) {
|
|
const int AUXBUF_LEN = 4096;
|
|
int to_read = ret;
|
|
float aux_buffer[AUXBUF_LEN];
|
|
|
|
while (to_read) {
|
|
int m = MIN(AUXBUF_LEN / vi.channels, to_read);
|
|
|
|
int count = 0;
|
|
|
|
for (int j = 0; j < m; j++) {
|
|
for (int i = 0; i < vi.channels; i++) {
|
|
aux_buffer[count++] = pcm[i][j];
|
|
}
|
|
}
|
|
|
|
if (mix_callback) {
|
|
int mixed = mix_callback(mix_udata, aux_buffer, m);
|
|
to_read -= mixed;
|
|
if (mixed != m) { //could mix no more
|
|
buffer_full = true;
|
|
break;
|
|
}
|
|
} else {
|
|
to_read -= m; //just pretend we sent the audio
|
|
}
|
|
}
|
|
|
|
vorbis_synthesis_read(&vd, ret - to_read);
|
|
|
|
audio_frames_wrote += ret - to_read;
|
|
|
|
} else {
|
|
/* no pending audio; is there a pending packet to decode? */
|
|
if (ogg_stream_packetout(&vo, &op) > 0) {
|
|
if (vorbis_synthesis(&vb, &op) == 0) { /* test for success! */
|
|
vorbis_synthesis_blockin(&vd, &vb);
|
|
}
|
|
} else { /* we need more data; break out to suck in another page */
|
|
break;
|
|
}
|
|
}
|
|
|
|
audio_done = videobuf_time < (audio_frames_wrote / float(vi.rate));
|
|
|
|
if (buffer_full) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
while (theora_p && !frame_done) {
|
|
/* theora is one in, one out... */
|
|
if (ogg_stream_packetout(&to, &op) > 0) {
|
|
/*HACK: This should be set after a seek or a gap, but we might not have
|
|
a granulepos for the first packet (we only have them for the last
|
|
packet on a page), so we just set it as often as we get it.
|
|
To do this right, we should back-track from the last packet on the
|
|
page and compute the correct granulepos for the first packet after
|
|
a seek or a gap.*/
|
|
if (op.granulepos >= 0) {
|
|
th_decode_ctl(td, TH_DECCTL_SET_GRANPOS, &op.granulepos,
|
|
sizeof(op.granulepos));
|
|
}
|
|
ogg_int64_t videobuf_granulepos;
|
|
if (th_decode_packetin(td, &op, &videobuf_granulepos) == 0) {
|
|
videobuf_time = th_granule_time(td, videobuf_granulepos);
|
|
|
|
//printf("frame time %f, play time %f, ready %i\n", (float)videobuf_time, get_time(), videobuf_ready);
|
|
|
|
/* is it already too old to be useful? This is only actually
|
|
useful cosmetically after a SIGSTOP. Note that we have to
|
|
decode the frame even if we don't show it (for now) due to
|
|
keyframing. Soon enough libtheora will be able to deal
|
|
with non-keyframe seeks. */
|
|
|
|
if (videobuf_time >= get_time()) {
|
|
frame_done = true;
|
|
} else {
|
|
/*If we are too slow, reduce the pp level.*/
|
|
pp_inc = pp_level > 0 ? -1 : 0;
|
|
}
|
|
}
|
|
|
|
} else {
|
|
no_theora = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
#ifdef THEORA_USE_THREAD_STREAMING
|
|
if (file.is_valid() && thread_eof && no_theora && theora_eos && ring_buffer.data_left() == 0) {
|
|
#else
|
|
if (file.is_valid() && /*!videobuf_ready && */ no_theora && theora_eos) {
|
|
#endif
|
|
//printf("video done, stopping\n");
|
|
stop();
|
|
return;
|
|
}
|
|
|
|
if (!frame_done || !audio_done) {
|
|
//what's the point of waiting for audio to grab a page?
|
|
|
|
buffer_data();
|
|
while (ogg_sync_pageout(&oy, &og) > 0) {
|
|
queue_page(&og);
|
|
}
|
|
}
|
|
|
|
/* If playback has begun, top audio buffer off immediately. */
|
|
//if(stateflag) audio_write_nonblocking();
|
|
|
|
/* are we at or past time for this video frame? */
|
|
if (videobuf_ready && videobuf_time <= get_time()) {
|
|
//video_write();
|
|
//videobuf_ready=0;
|
|
} else {
|
|
//printf("frame at %f not ready (time %f), ready %i\n", (float)videobuf_time, get_time(), videobuf_ready);
|
|
}
|
|
|
|
double tdiff = videobuf_time - get_time();
|
|
/*If we have lots of extra time, increase the post-processing level.*/
|
|
if (tdiff > ti.fps_denominator * 0.25 / ti.fps_numerator) {
|
|
pp_inc = pp_level < pp_level_max ? 1 : 0;
|
|
} else if (tdiff < ti.fps_denominator * 0.05 / ti.fps_numerator) {
|
|
pp_inc = pp_level > 0 ? -1 : 0;
|
|
}
|
|
}
|
|
|
|
video_write();
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::play() {
|
|
if (!playing) {
|
|
time = 0;
|
|
} else {
|
|
stop();
|
|
}
|
|
|
|
playing = true;
|
|
delay_compensation = GLOBAL_GET("audio/video/video_delay_compensation_ms");
|
|
delay_compensation /= 1000.0;
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::stop() {
|
|
if (playing) {
|
|
clear();
|
|
set_file(file_name); //reset
|
|
}
|
|
playing = false;
|
|
time = 0;
|
|
}
|
|
|
|
bool VideoStreamPlaybackTheora::is_playing() const {
|
|
return playing;
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::set_paused(bool p_paused) {
|
|
paused = p_paused;
|
|
}
|
|
|
|
bool VideoStreamPlaybackTheora::is_paused() const {
|
|
return paused;
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::set_loop(bool p_enable) {
|
|
}
|
|
|
|
bool VideoStreamPlaybackTheora::has_loop() const {
|
|
return false;
|
|
}
|
|
|
|
double VideoStreamPlaybackTheora::get_length() const {
|
|
return 0;
|
|
}
|
|
|
|
String VideoStreamPlaybackTheora::get_stream_name() const {
|
|
return "";
|
|
}
|
|
|
|
int VideoStreamPlaybackTheora::get_loop_count() const {
|
|
return 0;
|
|
}
|
|
|
|
double VideoStreamPlaybackTheora::get_playback_position() const {
|
|
return get_time();
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::seek(double p_time) {
|
|
WARN_PRINT_ONCE("Seeking in Theora videos is not implemented yet (it's only supported for GDExtension-provided video streams).");
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::set_mix_callback(AudioMixCallback p_callback, void *p_userdata) {
|
|
mix_callback = p_callback;
|
|
mix_udata = p_userdata;
|
|
}
|
|
|
|
int VideoStreamPlaybackTheora::get_channels() const {
|
|
return vi.channels;
|
|
}
|
|
|
|
void VideoStreamPlaybackTheora::set_audio_track(int p_idx) {
|
|
audio_track = p_idx;
|
|
}
|
|
|
|
int VideoStreamPlaybackTheora::get_mix_rate() const {
|
|
return vi.rate;
|
|
}
|
|
|
|
#ifdef THEORA_USE_THREAD_STREAMING
|
|
|
|
void VideoStreamPlaybackTheora::_streaming_thread(void *ud) {
|
|
VideoStreamPlaybackTheora *vs = static_cast<VideoStreamPlaybackTheora *>(ud);
|
|
|
|
while (!vs->thread_exit) {
|
|
//just fill back the buffer
|
|
if (!vs->thread_eof) {
|
|
int to_read = vs->ring_buffer.space_left();
|
|
if (to_read > 0) {
|
|
uint64_t read = vs->file->get_buffer(vs->read_buffer.ptr(), to_read);
|
|
vs->ring_buffer.write(vs->read_buffer.ptr(), read);
|
|
vs->thread_eof = vs->file->eof_reached();
|
|
}
|
|
}
|
|
|
|
vs->thread_sem->wait();
|
|
}
|
|
}
|
|
|
|
#endif
|
|
|
|
VideoStreamPlaybackTheora::VideoStreamPlaybackTheora() {
|
|
texture = Ref<ImageTexture>(memnew(ImageTexture));
|
|
|
|
#ifdef THEORA_USE_THREAD_STREAMING
|
|
int rb_power = nearest_shift(RB_SIZE_KB * 1024);
|
|
ring_buffer.resize(rb_power);
|
|
read_buffer.resize(RB_SIZE_KB * 1024);
|
|
thread_sem = Semaphore::create();
|
|
|
|
#endif
|
|
}
|
|
|
|
VideoStreamPlaybackTheora::~VideoStreamPlaybackTheora() {
|
|
#ifdef THEORA_USE_THREAD_STREAMING
|
|
memdelete(thread_sem);
|
|
#endif
|
|
clear();
|
|
}
|
|
|
|
void VideoStreamTheora::_bind_methods() {
|
|
ClassDB::bind_method(D_METHOD("set_file", "file"), &VideoStreamTheora::set_file);
|
|
ClassDB::bind_method(D_METHOD("get_file"), &VideoStreamTheora::get_file);
|
|
|
|
ADD_PROPERTY(PropertyInfo(Variant::STRING, "file", PROPERTY_HINT_NONE, "", PROPERTY_USAGE_NO_EDITOR | PROPERTY_USAGE_INTERNAL), "set_file", "get_file");
|
|
}
|
|
|
|
////////////
|
|
|
|
Ref<Resource> ResourceFormatLoaderTheora::load(const String &p_path, const String &p_original_path, Error *r_error, bool p_use_sub_threads, float *r_progress, CacheMode p_cache_mode) {
|
|
Ref<FileAccess> f = FileAccess::open(p_path, FileAccess::READ);
|
|
if (f.is_null()) {
|
|
if (r_error) {
|
|
*r_error = ERR_CANT_OPEN;
|
|
}
|
|
return Ref<Resource>();
|
|
}
|
|
|
|
VideoStreamTheora *stream = memnew(VideoStreamTheora);
|
|
stream->set_file(p_path);
|
|
|
|
Ref<VideoStreamTheora> ogv_stream = Ref<VideoStreamTheora>(stream);
|
|
|
|
if (r_error) {
|
|
*r_error = OK;
|
|
}
|
|
|
|
return ogv_stream;
|
|
}
|
|
|
|
void ResourceFormatLoaderTheora::get_recognized_extensions(List<String> *p_extensions) const {
|
|
p_extensions->push_back("ogv");
|
|
}
|
|
|
|
bool ResourceFormatLoaderTheora::handles_type(const String &p_type) const {
|
|
return ClassDB::is_parent_class(p_type, "VideoStream");
|
|
}
|
|
|
|
String ResourceFormatLoaderTheora::get_resource_type(const String &p_path) const {
|
|
String el = p_path.get_extension().to_lower();
|
|
if (el == "ogv") {
|
|
return "VideoStreamTheora";
|
|
}
|
|
return "";
|
|
}
|