mirror of
https://github.com/godotengine/godot.git
synced 2024-12-09 10:09:20 +08:00
d95794ec8a
As many open source projects have started doing it, we're removing the current year from the copyright notice, so that we don't need to bump it every year. It seems like only the first year of publication is technically relevant for copyright notices, and even that seems to be something that many companies stopped listing altogether (in a version controlled codebase, the commits are a much better source of date of publication than a hardcoded copyright statement). We also now list Godot Engine contributors first as we're collectively the current maintainers of the project, and we clarify that the "exclusive" copyright of the co-founders covers the timespan before opensourcing (their further contributions are included as part of Godot Engine contributors). Also fixed "cf." Frenchism - it's meant as "refer to / see".
662 lines
21 KiB
C++
662 lines
21 KiB
C++
/**************************************************************************/
|
|
/* http_request.cpp */
|
|
/**************************************************************************/
|
|
/* This file is part of: */
|
|
/* GODOT ENGINE */
|
|
/* https://godotengine.org */
|
|
/**************************************************************************/
|
|
/* Copyright (c) 2014-present Godot Engine contributors (see AUTHORS.md). */
|
|
/* Copyright (c) 2007-2014 Juan Linietsky, Ariel Manzur. */
|
|
/* */
|
|
/* Permission is hereby granted, free of charge, to any person obtaining */
|
|
/* a copy of this software and associated documentation files (the */
|
|
/* "Software"), to deal in the Software without restriction, including */
|
|
/* without limitation the rights to use, copy, modify, merge, publish, */
|
|
/* distribute, sublicense, and/or sell copies of the Software, and to */
|
|
/* permit persons to whom the Software is furnished to do so, subject to */
|
|
/* the following conditions: */
|
|
/* */
|
|
/* The above copyright notice and this permission notice shall be */
|
|
/* included in all copies or substantial portions of the Software. */
|
|
/* */
|
|
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
|
|
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
|
|
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. */
|
|
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
|
|
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
|
|
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
|
|
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
|
|
/**************************************************************************/
|
|
|
|
#include "http_request.h"
|
|
#include "core/io/compression.h"
|
|
#include "scene/main/timer.h"
|
|
|
|
Error HTTPRequest::_request() {
|
|
return client->connect_to_host(url, port, use_tls, validate_tls);
|
|
}
|
|
|
|
Error HTTPRequest::_parse_url(const String &p_url) {
|
|
use_tls = false;
|
|
request_string = "";
|
|
port = 80;
|
|
request_sent = false;
|
|
got_response = false;
|
|
body_len = -1;
|
|
body.clear();
|
|
downloaded.set(0);
|
|
final_body_size.set(0);
|
|
redirections = 0;
|
|
|
|
String scheme;
|
|
Error err = p_url.parse_url(scheme, url, port, request_string);
|
|
ERR_FAIL_COND_V_MSG(err != OK, err, "Error parsing URL: " + p_url + ".");
|
|
if (scheme == "https://") {
|
|
use_tls = true;
|
|
} else if (scheme != "http://") {
|
|
ERR_FAIL_V_MSG(ERR_INVALID_PARAMETER, "Invalid URL scheme: " + scheme + ".");
|
|
}
|
|
if (port == 0) {
|
|
port = use_tls ? 443 : 80;
|
|
}
|
|
if (request_string.is_empty()) {
|
|
request_string = "/";
|
|
}
|
|
return OK;
|
|
}
|
|
|
|
bool HTTPRequest::has_header(const PackedStringArray &p_headers, const String &p_header_name) {
|
|
bool exists = false;
|
|
|
|
String lower_case_header_name = p_header_name.to_lower();
|
|
for (int i = 0; i < p_headers.size() && !exists; i++) {
|
|
String sanitized = p_headers[i].strip_edges().to_lower();
|
|
if (sanitized.begins_with(lower_case_header_name)) {
|
|
exists = true;
|
|
}
|
|
}
|
|
|
|
return exists;
|
|
}
|
|
|
|
String HTTPRequest::get_header_value(const PackedStringArray &p_headers, const String &p_header_name) {
|
|
String value = "";
|
|
|
|
String lowwer_case_header_name = p_header_name.to_lower();
|
|
for (int i = 0; i < p_headers.size(); i++) {
|
|
if (p_headers[i].find(":") > 0) {
|
|
Vector<String> parts = p_headers[i].split(":", false, 1);
|
|
if (parts.size() > 1 && parts[0].strip_edges().to_lower() == lowwer_case_header_name) {
|
|
value = parts[1].strip_edges();
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
return value;
|
|
}
|
|
|
|
Error HTTPRequest::request(const String &p_url, const Vector<String> &p_custom_headers, bool p_tls_validate_domain, HTTPClient::Method p_method, const String &p_request_data) {
|
|
// Copy the string into a raw buffer.
|
|
Vector<uint8_t> raw_data;
|
|
|
|
CharString charstr = p_request_data.utf8();
|
|
size_t len = charstr.length();
|
|
if (len > 0) {
|
|
raw_data.resize(len);
|
|
uint8_t *w = raw_data.ptrw();
|
|
memcpy(w, charstr.ptr(), len);
|
|
}
|
|
|
|
return request_raw(p_url, p_custom_headers, p_tls_validate_domain, p_method, raw_data);
|
|
}
|
|
|
|
Error HTTPRequest::request_raw(const String &p_url, const Vector<String> &p_custom_headers, bool p_tls_validate_domain, HTTPClient::Method p_method, const Vector<uint8_t> &p_request_data_raw) {
|
|
ERR_FAIL_COND_V(!is_inside_tree(), ERR_UNCONFIGURED);
|
|
ERR_FAIL_COND_V_MSG(requesting, ERR_BUSY, "HTTPRequest is processing a request. Wait for completion or cancel it before attempting a new one.");
|
|
|
|
if (timeout > 0) {
|
|
timer->stop();
|
|
timer->start(timeout);
|
|
}
|
|
|
|
method = p_method;
|
|
|
|
Error err = _parse_url(p_url);
|
|
if (err) {
|
|
return err;
|
|
}
|
|
|
|
validate_tls = p_tls_validate_domain;
|
|
|
|
headers = p_custom_headers;
|
|
|
|
if (accept_gzip) {
|
|
// If the user has specified an Accept-Encoding header, don't overwrite it.
|
|
if (!has_header(headers, "Accept-Encoding")) {
|
|
headers.push_back("Accept-Encoding: gzip, deflate");
|
|
}
|
|
}
|
|
|
|
request_data = p_request_data_raw;
|
|
|
|
requesting = true;
|
|
|
|
if (use_threads.is_set()) {
|
|
thread_done.clear();
|
|
thread_request_quit.clear();
|
|
client->set_blocking_mode(true);
|
|
thread.start(_thread_func, this);
|
|
} else {
|
|
client->set_blocking_mode(false);
|
|
err = _request();
|
|
if (err != OK) {
|
|
_defer_done(RESULT_CANT_CONNECT, 0, PackedStringArray(), PackedByteArray());
|
|
return ERR_CANT_CONNECT;
|
|
}
|
|
|
|
set_process_internal(true);
|
|
}
|
|
|
|
return OK;
|
|
}
|
|
|
|
void HTTPRequest::_thread_func(void *p_userdata) {
|
|
HTTPRequest *hr = static_cast<HTTPRequest *>(p_userdata);
|
|
|
|
Error err = hr->_request();
|
|
|
|
if (err != OK) {
|
|
hr->_defer_done(RESULT_CANT_CONNECT, 0, PackedStringArray(), PackedByteArray());
|
|
} else {
|
|
while (!hr->thread_request_quit.is_set()) {
|
|
bool exit = hr->_update_connection();
|
|
if (exit) {
|
|
break;
|
|
}
|
|
OS::get_singleton()->delay_usec(1);
|
|
}
|
|
}
|
|
|
|
hr->thread_done.set();
|
|
}
|
|
|
|
void HTTPRequest::cancel_request() {
|
|
timer->stop();
|
|
|
|
if (!requesting) {
|
|
return;
|
|
}
|
|
|
|
if (!use_threads.is_set()) {
|
|
set_process_internal(false);
|
|
} else {
|
|
thread_request_quit.set();
|
|
thread.wait_to_finish();
|
|
}
|
|
|
|
file.unref();
|
|
decompressor.unref();
|
|
client->close();
|
|
body.clear();
|
|
got_response = false;
|
|
response_code = -1;
|
|
request_sent = false;
|
|
requesting = false;
|
|
}
|
|
|
|
bool HTTPRequest::_handle_response(bool *ret_value) {
|
|
if (!client->has_response()) {
|
|
_defer_done(RESULT_NO_RESPONSE, 0, PackedStringArray(), PackedByteArray());
|
|
*ret_value = true;
|
|
return true;
|
|
}
|
|
|
|
got_response = true;
|
|
response_code = client->get_response_code();
|
|
List<String> rheaders;
|
|
client->get_response_headers(&rheaders);
|
|
response_headers.clear();
|
|
downloaded.set(0);
|
|
final_body_size.set(0);
|
|
decompressor.unref();
|
|
|
|
for (const String &E : rheaders) {
|
|
response_headers.push_back(E);
|
|
}
|
|
|
|
if (response_code == 301 || response_code == 302) {
|
|
// Handle redirect.
|
|
|
|
if (max_redirects >= 0 && redirections >= max_redirects) {
|
|
_defer_done(RESULT_REDIRECT_LIMIT_REACHED, response_code, response_headers, PackedByteArray());
|
|
*ret_value = true;
|
|
return true;
|
|
}
|
|
|
|
String new_request;
|
|
|
|
for (const String &E : rheaders) {
|
|
if (E.findn("Location: ") != -1) {
|
|
new_request = E.substr(9, E.length()).strip_edges();
|
|
}
|
|
}
|
|
|
|
if (!new_request.is_empty()) {
|
|
// Process redirect.
|
|
client->close();
|
|
int new_redirs = redirections + 1; // Because _request() will clear it.
|
|
Error err;
|
|
if (new_request.begins_with("http")) {
|
|
// New url, new request.
|
|
_parse_url(new_request);
|
|
} else {
|
|
request_string = new_request;
|
|
}
|
|
|
|
err = _request();
|
|
if (err == OK) {
|
|
request_sent = false;
|
|
got_response = false;
|
|
body_len = -1;
|
|
body.clear();
|
|
downloaded.set(0);
|
|
final_body_size.set(0);
|
|
redirections = new_redirs;
|
|
*ret_value = false;
|
|
return true;
|
|
}
|
|
}
|
|
}
|
|
|
|
// Check if we need to start streaming decompression.
|
|
String content_encoding;
|
|
if (accept_gzip) {
|
|
content_encoding = get_header_value(response_headers, "Content-Encoding").to_lower();
|
|
}
|
|
if (content_encoding == "gzip") {
|
|
decompressor.instantiate();
|
|
decompressor->start_decompression(false, get_download_chunk_size());
|
|
} else if (content_encoding == "deflate") {
|
|
decompressor.instantiate();
|
|
decompressor->start_decompression(true, get_download_chunk_size());
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool HTTPRequest::_update_connection() {
|
|
switch (client->get_status()) {
|
|
case HTTPClient::STATUS_DISCONNECTED: {
|
|
_defer_done(RESULT_CANT_CONNECT, 0, PackedStringArray(), PackedByteArray());
|
|
return true; // End it, since it's disconnected.
|
|
} break;
|
|
case HTTPClient::STATUS_RESOLVING: {
|
|
client->poll();
|
|
// Must wait.
|
|
return false;
|
|
} break;
|
|
case HTTPClient::STATUS_CANT_RESOLVE: {
|
|
_defer_done(RESULT_CANT_RESOLVE, 0, PackedStringArray(), PackedByteArray());
|
|
return true;
|
|
|
|
} break;
|
|
case HTTPClient::STATUS_CONNECTING: {
|
|
client->poll();
|
|
// Must wait.
|
|
return false;
|
|
} break; // Connecting to IP.
|
|
case HTTPClient::STATUS_CANT_CONNECT: {
|
|
_defer_done(RESULT_CANT_CONNECT, 0, PackedStringArray(), PackedByteArray());
|
|
return true;
|
|
|
|
} break;
|
|
case HTTPClient::STATUS_CONNECTED: {
|
|
if (request_sent) {
|
|
if (!got_response) {
|
|
// No body.
|
|
|
|
bool ret_value;
|
|
|
|
if (_handle_response(&ret_value)) {
|
|
return ret_value;
|
|
}
|
|
|
|
_defer_done(RESULT_SUCCESS, response_code, response_headers, PackedByteArray());
|
|
return true;
|
|
}
|
|
if (body_len < 0) {
|
|
// Chunked transfer is done.
|
|
_defer_done(RESULT_SUCCESS, response_code, response_headers, body);
|
|
return true;
|
|
}
|
|
|
|
_defer_done(RESULT_CHUNKED_BODY_SIZE_MISMATCH, response_code, response_headers, PackedByteArray());
|
|
return true;
|
|
// Request might have been done.
|
|
} else {
|
|
// Did not request yet, do request.
|
|
|
|
int size = request_data.size();
|
|
Error err = client->request(method, request_string, headers, size > 0 ? request_data.ptr() : nullptr, size);
|
|
if (err != OK) {
|
|
_defer_done(RESULT_CONNECTION_ERROR, 0, PackedStringArray(), PackedByteArray());
|
|
return true;
|
|
}
|
|
|
|
request_sent = true;
|
|
return false;
|
|
}
|
|
} break; // Connected: break requests only accepted here.
|
|
case HTTPClient::STATUS_REQUESTING: {
|
|
// Must wait, still requesting.
|
|
client->poll();
|
|
return false;
|
|
|
|
} break; // Request in progress.
|
|
case HTTPClient::STATUS_BODY: {
|
|
if (!got_response) {
|
|
bool ret_value;
|
|
|
|
if (_handle_response(&ret_value)) {
|
|
return ret_value;
|
|
}
|
|
|
|
if (!client->is_response_chunked() && client->get_response_body_length() == 0) {
|
|
_defer_done(RESULT_SUCCESS, response_code, response_headers, PackedByteArray());
|
|
return true;
|
|
}
|
|
|
|
// No body len (-1) if chunked or no content-length header was provided.
|
|
// Change your webserver configuration if you want body len.
|
|
body_len = client->get_response_body_length();
|
|
|
|
if (body_size_limit >= 0 && body_len > body_size_limit) {
|
|
_defer_done(RESULT_BODY_SIZE_LIMIT_EXCEEDED, response_code, response_headers, PackedByteArray());
|
|
return true;
|
|
}
|
|
|
|
if (!download_to_file.is_empty()) {
|
|
file = FileAccess::open(download_to_file, FileAccess::WRITE);
|
|
if (file.is_null()) {
|
|
_defer_done(RESULT_DOWNLOAD_FILE_CANT_OPEN, response_code, response_headers, PackedByteArray());
|
|
return true;
|
|
}
|
|
}
|
|
}
|
|
|
|
client->poll();
|
|
if (client->get_status() != HTTPClient::STATUS_BODY) {
|
|
return false;
|
|
}
|
|
|
|
PackedByteArray chunk;
|
|
if (decompressor.is_null()) {
|
|
// Chunk can be read directly.
|
|
chunk = client->read_response_body_chunk();
|
|
downloaded.add(chunk.size());
|
|
} else {
|
|
// Chunk is the result of decompression.
|
|
PackedByteArray compressed = client->read_response_body_chunk();
|
|
downloaded.add(compressed.size());
|
|
|
|
int pos = 0;
|
|
int left = compressed.size();
|
|
while (left) {
|
|
int w = 0;
|
|
Error err = decompressor->put_partial_data(compressed.ptr() + pos, left, w);
|
|
if (err == OK) {
|
|
PackedByteArray dc;
|
|
dc.resize(decompressor->get_available_bytes());
|
|
err = decompressor->get_data(dc.ptrw(), dc.size());
|
|
chunk.append_array(dc);
|
|
}
|
|
if (err != OK) {
|
|
_defer_done(RESULT_BODY_DECOMPRESS_FAILED, response_code, response_headers, PackedByteArray());
|
|
return true;
|
|
}
|
|
// We need this check here because a "zip bomb" could result in a chunk of few kilos decompressing into gigabytes of data.
|
|
if (body_size_limit >= 0 && final_body_size.get() + chunk.size() > body_size_limit) {
|
|
_defer_done(RESULT_BODY_SIZE_LIMIT_EXCEEDED, response_code, response_headers, PackedByteArray());
|
|
return true;
|
|
}
|
|
pos += w;
|
|
left -= w;
|
|
}
|
|
}
|
|
final_body_size.add(chunk.size());
|
|
|
|
if (body_size_limit >= 0 && final_body_size.get() > body_size_limit) {
|
|
_defer_done(RESULT_BODY_SIZE_LIMIT_EXCEEDED, response_code, response_headers, PackedByteArray());
|
|
return true;
|
|
}
|
|
|
|
if (chunk.size()) {
|
|
if (file.is_valid()) {
|
|
const uint8_t *r = chunk.ptr();
|
|
file->store_buffer(r, chunk.size());
|
|
if (file->get_error() != OK) {
|
|
_defer_done(RESULT_DOWNLOAD_FILE_WRITE_ERROR, response_code, response_headers, PackedByteArray());
|
|
return true;
|
|
}
|
|
} else {
|
|
body.append_array(chunk);
|
|
}
|
|
}
|
|
|
|
if (body_len >= 0) {
|
|
if (downloaded.get() == body_len) {
|
|
_defer_done(RESULT_SUCCESS, response_code, response_headers, body);
|
|
return true;
|
|
}
|
|
} else if (client->get_status() == HTTPClient::STATUS_DISCONNECTED) {
|
|
// We read till EOF, with no errors. Request is done.
|
|
_defer_done(RESULT_SUCCESS, response_code, response_headers, body);
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
|
|
} break; // Request resulted in body: break which must be read.
|
|
case HTTPClient::STATUS_CONNECTION_ERROR: {
|
|
_defer_done(RESULT_CONNECTION_ERROR, 0, PackedStringArray(), PackedByteArray());
|
|
return true;
|
|
} break;
|
|
case HTTPClient::STATUS_TLS_HANDSHAKE_ERROR: {
|
|
_defer_done(RESULT_TLS_HANDSHAKE_ERROR, 0, PackedStringArray(), PackedByteArray());
|
|
return true;
|
|
} break;
|
|
}
|
|
|
|
ERR_FAIL_V(false);
|
|
}
|
|
|
|
void HTTPRequest::_defer_done(int p_status, int p_code, const PackedStringArray &p_headers, const PackedByteArray &p_data) {
|
|
call_deferred(SNAME("_request_done"), p_status, p_code, p_headers, p_data);
|
|
}
|
|
|
|
void HTTPRequest::_request_done(int p_status, int p_code, const PackedStringArray &p_headers, const PackedByteArray &p_data) {
|
|
cancel_request();
|
|
|
|
emit_signal(SNAME("request_completed"), p_status, p_code, p_headers, p_data);
|
|
}
|
|
|
|
void HTTPRequest::_notification(int p_what) {
|
|
switch (p_what) {
|
|
case NOTIFICATION_INTERNAL_PROCESS: {
|
|
if (use_threads.is_set()) {
|
|
return;
|
|
}
|
|
bool done = _update_connection();
|
|
if (done) {
|
|
set_process_internal(false);
|
|
}
|
|
} break;
|
|
|
|
case NOTIFICATION_EXIT_TREE: {
|
|
if (requesting) {
|
|
cancel_request();
|
|
}
|
|
} break;
|
|
}
|
|
}
|
|
|
|
void HTTPRequest::set_use_threads(bool p_use) {
|
|
ERR_FAIL_COND(get_http_client_status() != HTTPClient::STATUS_DISCONNECTED);
|
|
use_threads.set_to(p_use);
|
|
}
|
|
|
|
bool HTTPRequest::is_using_threads() const {
|
|
return use_threads.is_set();
|
|
}
|
|
|
|
void HTTPRequest::set_accept_gzip(bool p_gzip) {
|
|
accept_gzip = p_gzip;
|
|
}
|
|
|
|
bool HTTPRequest::is_accepting_gzip() const {
|
|
return accept_gzip;
|
|
}
|
|
|
|
void HTTPRequest::set_body_size_limit(int p_bytes) {
|
|
ERR_FAIL_COND(get_http_client_status() != HTTPClient::STATUS_DISCONNECTED);
|
|
|
|
body_size_limit = p_bytes;
|
|
}
|
|
|
|
int HTTPRequest::get_body_size_limit() const {
|
|
return body_size_limit;
|
|
}
|
|
|
|
void HTTPRequest::set_download_file(const String &p_file) {
|
|
ERR_FAIL_COND(get_http_client_status() != HTTPClient::STATUS_DISCONNECTED);
|
|
|
|
download_to_file = p_file;
|
|
}
|
|
|
|
String HTTPRequest::get_download_file() const {
|
|
return download_to_file;
|
|
}
|
|
|
|
void HTTPRequest::set_download_chunk_size(int p_chunk_size) {
|
|
ERR_FAIL_COND(get_http_client_status() != HTTPClient::STATUS_DISCONNECTED);
|
|
|
|
client->set_read_chunk_size(p_chunk_size);
|
|
}
|
|
|
|
int HTTPRequest::get_download_chunk_size() const {
|
|
return client->get_read_chunk_size();
|
|
}
|
|
|
|
HTTPClient::Status HTTPRequest::get_http_client_status() const {
|
|
return client->get_status();
|
|
}
|
|
|
|
void HTTPRequest::set_max_redirects(int p_max) {
|
|
max_redirects = p_max;
|
|
}
|
|
|
|
int HTTPRequest::get_max_redirects() const {
|
|
return max_redirects;
|
|
}
|
|
|
|
int HTTPRequest::get_downloaded_bytes() const {
|
|
return downloaded.get();
|
|
}
|
|
|
|
int HTTPRequest::get_body_size() const {
|
|
return body_len;
|
|
}
|
|
|
|
void HTTPRequest::set_http_proxy(const String &p_host, int p_port) {
|
|
client->set_http_proxy(p_host, p_port);
|
|
}
|
|
|
|
void HTTPRequest::set_https_proxy(const String &p_host, int p_port) {
|
|
client->set_https_proxy(p_host, p_port);
|
|
}
|
|
|
|
void HTTPRequest::set_timeout(double p_timeout) {
|
|
ERR_FAIL_COND(p_timeout < 0);
|
|
timeout = p_timeout;
|
|
}
|
|
|
|
double HTTPRequest::get_timeout() {
|
|
return timeout;
|
|
}
|
|
|
|
void HTTPRequest::_timeout() {
|
|
cancel_request();
|
|
_defer_done(RESULT_TIMEOUT, 0, PackedStringArray(), PackedByteArray());
|
|
}
|
|
|
|
void HTTPRequest::_bind_methods() {
|
|
ClassDB::bind_method(D_METHOD("request", "url", "custom_headers", "tls_validate_domain", "method", "request_data"), &HTTPRequest::request, DEFVAL(PackedStringArray()), DEFVAL(true), DEFVAL(HTTPClient::METHOD_GET), DEFVAL(String()));
|
|
ClassDB::bind_method(D_METHOD("request_raw", "url", "custom_headers", "tls_validate_domain", "method", "request_data_raw"), &HTTPRequest::request_raw, DEFVAL(PackedStringArray()), DEFVAL(true), DEFVAL(HTTPClient::METHOD_GET), DEFVAL(PackedByteArray()));
|
|
ClassDB::bind_method(D_METHOD("cancel_request"), &HTTPRequest::cancel_request);
|
|
|
|
ClassDB::bind_method(D_METHOD("get_http_client_status"), &HTTPRequest::get_http_client_status);
|
|
|
|
ClassDB::bind_method(D_METHOD("set_use_threads", "enable"), &HTTPRequest::set_use_threads);
|
|
ClassDB::bind_method(D_METHOD("is_using_threads"), &HTTPRequest::is_using_threads);
|
|
|
|
ClassDB::bind_method(D_METHOD("set_accept_gzip", "enable"), &HTTPRequest::set_accept_gzip);
|
|
ClassDB::bind_method(D_METHOD("is_accepting_gzip"), &HTTPRequest::is_accepting_gzip);
|
|
|
|
ClassDB::bind_method(D_METHOD("set_body_size_limit", "bytes"), &HTTPRequest::set_body_size_limit);
|
|
ClassDB::bind_method(D_METHOD("get_body_size_limit"), &HTTPRequest::get_body_size_limit);
|
|
|
|
ClassDB::bind_method(D_METHOD("set_max_redirects", "amount"), &HTTPRequest::set_max_redirects);
|
|
ClassDB::bind_method(D_METHOD("get_max_redirects"), &HTTPRequest::get_max_redirects);
|
|
|
|
ClassDB::bind_method(D_METHOD("set_download_file", "path"), &HTTPRequest::set_download_file);
|
|
ClassDB::bind_method(D_METHOD("get_download_file"), &HTTPRequest::get_download_file);
|
|
|
|
ClassDB::bind_method(D_METHOD("get_downloaded_bytes"), &HTTPRequest::get_downloaded_bytes);
|
|
ClassDB::bind_method(D_METHOD("get_body_size"), &HTTPRequest::get_body_size);
|
|
|
|
ClassDB::bind_method(D_METHOD("_request_done"), &HTTPRequest::_request_done);
|
|
|
|
ClassDB::bind_method(D_METHOD("set_timeout", "timeout"), &HTTPRequest::set_timeout);
|
|
ClassDB::bind_method(D_METHOD("get_timeout"), &HTTPRequest::get_timeout);
|
|
|
|
ClassDB::bind_method(D_METHOD("set_download_chunk_size", "chunk_size"), &HTTPRequest::set_download_chunk_size);
|
|
ClassDB::bind_method(D_METHOD("get_download_chunk_size"), &HTTPRequest::get_download_chunk_size);
|
|
|
|
ClassDB::bind_method(D_METHOD("set_http_proxy", "host", "port"), &HTTPRequest::set_http_proxy);
|
|
ClassDB::bind_method(D_METHOD("set_https_proxy", "host", "port"), &HTTPRequest::set_https_proxy);
|
|
|
|
ADD_PROPERTY(PropertyInfo(Variant::STRING, "download_file", PROPERTY_HINT_FILE), "set_download_file", "get_download_file");
|
|
ADD_PROPERTY(PropertyInfo(Variant::INT, "download_chunk_size", PROPERTY_HINT_RANGE, "256,16777216,suffix:B"), "set_download_chunk_size", "get_download_chunk_size");
|
|
ADD_PROPERTY(PropertyInfo(Variant::BOOL, "use_threads"), "set_use_threads", "is_using_threads");
|
|
ADD_PROPERTY(PropertyInfo(Variant::BOOL, "accept_gzip"), "set_accept_gzip", "is_accepting_gzip");
|
|
ADD_PROPERTY(PropertyInfo(Variant::INT, "body_size_limit", PROPERTY_HINT_RANGE, "-1,2000000000,suffix:B"), "set_body_size_limit", "get_body_size_limit");
|
|
ADD_PROPERTY(PropertyInfo(Variant::INT, "max_redirects", PROPERTY_HINT_RANGE, "-1,64"), "set_max_redirects", "get_max_redirects");
|
|
ADD_PROPERTY(PropertyInfo(Variant::FLOAT, "timeout", PROPERTY_HINT_RANGE, "0,3600,0.1,or_greater,suffix:s"), "set_timeout", "get_timeout");
|
|
|
|
ADD_SIGNAL(MethodInfo("request_completed", PropertyInfo(Variant::INT, "result"), PropertyInfo(Variant::INT, "response_code"), PropertyInfo(Variant::PACKED_STRING_ARRAY, "headers"), PropertyInfo(Variant::PACKED_BYTE_ARRAY, "body")));
|
|
|
|
BIND_ENUM_CONSTANT(RESULT_SUCCESS);
|
|
BIND_ENUM_CONSTANT(RESULT_CHUNKED_BODY_SIZE_MISMATCH);
|
|
BIND_ENUM_CONSTANT(RESULT_CANT_CONNECT);
|
|
BIND_ENUM_CONSTANT(RESULT_CANT_RESOLVE);
|
|
BIND_ENUM_CONSTANT(RESULT_CONNECTION_ERROR);
|
|
BIND_ENUM_CONSTANT(RESULT_TLS_HANDSHAKE_ERROR);
|
|
BIND_ENUM_CONSTANT(RESULT_NO_RESPONSE);
|
|
BIND_ENUM_CONSTANT(RESULT_BODY_SIZE_LIMIT_EXCEEDED);
|
|
BIND_ENUM_CONSTANT(RESULT_BODY_DECOMPRESS_FAILED);
|
|
BIND_ENUM_CONSTANT(RESULT_REQUEST_FAILED);
|
|
BIND_ENUM_CONSTANT(RESULT_DOWNLOAD_FILE_CANT_OPEN);
|
|
BIND_ENUM_CONSTANT(RESULT_DOWNLOAD_FILE_WRITE_ERROR);
|
|
BIND_ENUM_CONSTANT(RESULT_REDIRECT_LIMIT_REACHED);
|
|
BIND_ENUM_CONSTANT(RESULT_TIMEOUT);
|
|
}
|
|
|
|
HTTPRequest::HTTPRequest() {
|
|
client = Ref<HTTPClient>(HTTPClient::create());
|
|
timer = memnew(Timer);
|
|
timer->set_one_shot(true);
|
|
timer->connect("timeout", callable_mp(this, &HTTPRequest::_timeout));
|
|
add_child(timer);
|
|
}
|