X-Git-Url: http://git.wpitchoune.net/gitweb/?a=blobdiff_plain;f=src%2Flp_ws.c;h=81b8b63086c01472c677a612aed11d6dd4d78bba;hb=49b52ead6d48488a4c1f90f372564c834da803cb;hp=6c2e299b18110dc618fd9e4822c0c3ca3e8af4a7;hpb=cacd81cd1065cf9a145bb2fbe66c383ad3c90206;p=ppastats.git diff --git a/src/lp_ws.c b/src/lp_ws.c index 6c2e299..81b8b63 100644 --- a/src/lp_ws.c +++ b/src/lp_ws.c @@ -16,24 +16,29 @@ Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ +#include +#define _(String) gettext(String) #include #include +#include #include #include #include "cache.h" #include "list.h" +#include "log.h" #include "lp_ws.h" #include "lp_json.h" #include "ppastats.h" -#define QUERY_GET_PUBLISHED_BINARIES \ - "?ws.op=getPublishedBinaries" -#define QUERY_GET_DOWNLOAD_COUNT "?ws.op=getDownloadCount" -#define QUERY_GET_DAILY_DOWNLOAD_TOTALS \ - "?ws.op=getDailyDownloadTotals" +static const char *QUERY_GET_PUBLISHED_BINARIES = "?ws.op=getPublishedBinaries"; +static const char *QUERY_GET_DOWNLOAD_COUNT = "?ws.op=getDownloadCount"; +static const char * +QUERY_GET_DAILY_DOWNLOAD_TOTALS = "?ws.op=getDailyDownloadTotals"; + +static const int DEFAULT_FETCH_RETRIES = 3; static CURL *curl; @@ -59,13 +64,14 @@ static size_t cbk_curl(void *buffer, size_t size, size_t nmemb, void *userp) static char *fetch_url(const char *url) { struct ucontent *content = malloc(sizeof(struct ucontent)); - char *result = NULL; + char *result; long code; + int retries; - if (debug) - printf("DEBUG: fetch_url %s\n", url); + log_debug(_("fetch_url(): %s"), url); if (!curl) { + log_debug(_("initializing CURL")); curl_global_init(CURL_GLOBAL_ALL); curl = curl_easy_init(); } @@ -73,6 +79,11 @@ static char *fetch_url(const char *url) if (!curl) exit(EXIT_FAILURE); + result = NULL; + + retries = DEFAULT_FETCH_RETRIES; + + retrieve: content->data = malloc(1); content->data[0] = '\0'; content->len = 0; @@ -84,10 +95,32 @@ static char *fetch_url(const char *url) curl_easy_setopt(curl, CURLOPT_USERAGENT, "ppastats/0.0"); if (curl_easy_perform(curl) == CURLE_OK) { - curl_easy_getinfo(curl, CURLINFO_RESPONSE_CODE, &code); - if (code == 200) + + switch (code) { + case 200: result = content->data; + break; + case 500: + case 502: + case 503: + case 504: + if (retries) { + log_err(_("Fetch failed with code %ld " + "for URL %s"), + code, + url); + + log_debug(_("Wait 5s before retry")); + sleep(5); + + free(content->data); + retries--; + goto retrieve; + } + default: + log_err(_("Fetch failed: %ld"), code); + } } if (!result) @@ -121,23 +154,24 @@ json_object_to_binary_package_publishing_history_list struct binary_package_publishing_history * * get_binary_package_publishing_history_list(const char *archive_url, - const char *package_status) + const char *pkg_status) { struct json_object *o_next; - char *url = malloc(strlen(archive_url)+ - strlen(QUERY_GET_PUBLISHED_BINARIES)+ - strlen("&status=")+ - 9+ - 1); + char *url; json_object *o; void **result = NULL; + url = malloc(strlen(archive_url)+ + strlen(QUERY_GET_PUBLISHED_BINARIES)+ + (pkg_status ? strlen("&status=")+strlen(pkg_status) : 0)+ + 1); + strcpy(url, archive_url); strcat(url, QUERY_GET_PUBLISHED_BINARIES); - if (package_status) { + if (pkg_status) { strcat(url, "&status="); - strcat(url, package_status); + strcat(url, pkg_status); } while (url) { @@ -148,8 +182,8 @@ get_binary_package_publishing_history_list(const char *archive_url, if (!o) break; - result = list_add_list(result, - (void **)json_object_to_bpph_list(o)); + result = list_append_list(result, + (void **)json_object_to_bpph_list(o)); o_next = json_object_object_get(o, "next_collection_link"); @@ -257,8 +291,7 @@ struct daily_download_total **get_daily_download_totals(const char *binary_url) void lp_ws_cleanup() { - if (debug) - printf("DEBUG: cleanup CURL\n"); + log_debug(_("cleanup CURL")); curl_easy_cleanup(curl); curl_global_cleanup();