X-Git-Url: http://git.wpitchoune.net/gitweb/?a=blobdiff_plain;f=src%2Flp_ws.c;h=ecc95cc30c7c26fd87713ea27f02be2c2ce5782d;hb=23b63657e6e51b811ff70824a69c594039cfbf66;hp=023ba9b28f17e58af7f096adf0bea4fac4edbaad;hpb=be68914bbcbfde6dbc51e20075691afbab3c2489;p=ppastats.git diff --git a/src/lp_ws.c b/src/lp_ws.c index 023ba9b..ecc95cc 100644 --- a/src/lp_ws.c +++ b/src/lp_ws.c @@ -20,15 +20,15 @@ #include #define _(String) gettext(String) +#include #include #include -#include -#include #include #include "cache.h" #include "fcache.h" +#include "http.h" #include "list.h" #include "log.h" #include "lp_ws.h" @@ -36,167 +36,133 @@ #include "ppastats.h" static const char * -QUERY_GET_PUBLISHED_BINARIES = "?ws.op=getPublishedBinaries&ws.size=300"; +QUERY_GET_PUBLISHED_BINARIES = "?ws.op=getPublishedBinaries&ws.size=150"; static const char *QUERY_GET_DOWNLOAD_COUNT = "?ws.op=getDownloadCount"; static const char * QUERY_GET_DAILY_DOWNLOAD_TOTALS = "?ws.op=getDailyDownloadTotals"; -static const int DEFAULT_FETCH_RETRIES = 10; - -static CURL *curl; - -struct ucontent { - char *data; - size_t len; -}; - -static size_t cbk_curl(void *buffer, size_t size, size_t nmemb, void *userp) +static json_object *get_json_object(const char *url) { - size_t realsize = size * nmemb; - struct ucontent *mem = (struct ucontent *)userp; + json_object *obj = NULL; + char *body; - mem->data = realloc(mem->data, mem->len + realsize + 1); + body = get_url_content(url, 0); - memcpy(&(mem->data[mem->len]), buffer, realsize); - mem->len += realsize; - mem->data[mem->len] = 0; + if (body) { + obj = json_tokener_parse(body); - return realsize; -} + free(body); -static void init() -{ - if (!curl) { - log_debug(_("initializing CURL")); - curl_global_init(CURL_GLOBAL_ALL); - curl = curl_easy_init(); + return obj; } - if (!curl) - exit(EXIT_FAILURE); + return NULL; } -static char *fetch_url(const char *url) +static char *get_bpph_list_cache_key(const char *archive_url) { - struct ucontent *content = malloc(sizeof(struct ucontent)); - char *result; - long code; - int retries; - unsigned int s; - - log_debug(_("fetch_url(): %s"), url); - - init(); - - result = NULL; - - retries = DEFAULT_FETCH_RETRIES; + char *key; - retrieve: - content->data = malloc(1); - content->data[0] = '\0'; - content->len = 0; + key = malloc(strlen(archive_url + 7) + strlen("/bpph") + 1); + sprintf(key, "%s/bpph", archive_url + 7); - curl_easy_setopt(curl, CURLOPT_URL, url); - curl_easy_setopt(curl, CURLOPT_VERBOSE, 0); - curl_easy_setopt(curl, CURLOPT_WRITEFUNCTION, cbk_curl); - curl_easy_setopt(curl, CURLOPT_WRITEDATA, content); - curl_easy_setopt(curl, CURLOPT_USERAGENT, "ppastats/0.0"); + return key; +} - if (curl_easy_perform(curl) == CURLE_OK) { - curl_easy_getinfo(curl, CURLINFO_RESPONSE_CODE, &code); +static struct bpph **get_bpph_list_from_cache(const char *key) +{ + char *content; + struct bpph **list; + json_object *json; - switch (code) { - case 200: - result = content->data; - break; - case 500: - case 502: - case 503: - case 504: - log_err(_("Fetch failed with code %ld for URL %s"), - code, - url); - - if (retries) { - s = 2 * (DEFAULT_FETCH_RETRIES - retries) + 2; - log_debug(_("Wait %ds before retry"), s); - sleep(s); - - free(content->data); - retries--; - goto retrieve; - } + content = fcache_get(key); + if (!content) + return NULL; - break; - default: - log_err(_("Fetch failed with code %ld for URL %s"), - code, - url); - } - } + json = json_tokener_parse(content); + if (!json) + return NULL; - if (!result) - free(content->data); + list = json_object_to_bpph_list(json); + json_object_put(json); free(content); - return result; + return list; } -static json_object *get_json_object(const char *url) +static char *get_last_creation_date(struct bpph **list) { - json_object *obj = NULL; - char *body; + time_t last, t; + struct bpph **cur; - body = fetch_url(url); + last = 0; - if (body) { - obj = json_tokener_parse(body); + if (list) + for (cur = list; *cur; cur++) { + t = (*cur)->date_created; + if (t > last) + last = t; + } - free(body); + if (last) + return time_to_str(last); + else + return NULL; +} - return obj; - } +struct bpph **get_bpph_list(const char *archive_url, const char *pkg_status) +{ + char *url, *key, *tmp; + struct bpph **result; + struct json_object *o, *bpph_json, *o_next; + char *created_since_date; + int ok; + + url = malloc(strlen(archive_url) + + strlen(QUERY_GET_PUBLISHED_BINARIES) + + 1); + strcpy(url, archive_url); + strcat(url, QUERY_GET_PUBLISHED_BINARIES); - return NULL; -} + key = get_bpph_list_cache_key(archive_url); -#define json_object_to_bpph_list \ -json_object_to_binary_package_publishing_history_list + result = get_bpph_list_from_cache(key); -struct binary_package_publishing_history * * -get_binary_package_publishing_history_list(const char *archive_url, - const char *pkg_status) -{ - struct json_object *o_next; - char *url; - json_object *o; - void **result = NULL; + if (result) { + created_since_date = get_last_creation_date(result); - url = malloc(strlen(archive_url)+ - strlen(QUERY_GET_PUBLISHED_BINARIES)+ - (pkg_status ? strlen("&status=")+strlen(pkg_status) : 0)+ - 1); + if (created_since_date) { + printf("Update package since: %s\n", + created_since_date); - strcpy(url, archive_url); - strcat(url, QUERY_GET_PUBLISHED_BINARIES); + tmp = malloc(strlen(url) + + strlen("&created_since_date=") + + strlen(created_since_date)+1); + strcpy(tmp, url); + strcat(tmp, "&created_since_date="); + strcat(tmp, created_since_date); - if (pkg_status) { - strcat(url, "&status="); - strcat(url, pkg_status); + free(url); + url = tmp; + + free(created_since_date); + } } + ok = 1; while (url) { o = get_json_object(url); free(url); url = NULL; - if (!o) + if (!o) { + ok = 0; break; + } - result = list_append_list(result, - (void **)json_object_to_bpph_list(o)); + result = bpph_list_append_list(result, + json_object_to_bpph_list(o)); o_next = json_object_object_get(o, "next_collection_link"); @@ -206,7 +172,15 @@ get_binary_package_publishing_history_list(const char *archive_url, json_object_put(o); } - return (struct binary_package_publishing_history **)result; + if (ok) { + bpph_json = bpph_list_to_json(result); + fcache_put(key, json_object_to_json_string(bpph_json)); + json_object_put(bpph_json); + } + + free(key); + + return result; } int get_download_count(const char *archive_url) @@ -242,14 +216,10 @@ const struct distro_arch_series *get_distro_arch_series(const char *url) if (distro) return (struct distro_arch_series *)distro; - content = fcache_get(url + 7); - if (!content) { - content = fetch_url(url); - if (content) - fcache_put(url + 7, content); - else - return NULL; - } + content = get_url_content(url, 1); + + if (!content) + return NULL; obj = json_tokener_parse(content); @@ -271,12 +241,20 @@ const struct distro_series *get_distro_series(const char *url) { json_object *obj; const struct distro_series *distro; + char *content; distro = cache_get(url); if (distro) return (struct distro_series *)distro; - obj = get_json_object(url); + content = get_url_content(url, 1); + + if (!content) + return NULL; + + obj = json_tokener_parse(content); + + free(content); if (!obj) return NULL; @@ -314,10 +292,3 @@ struct daily_download_total **get_daily_download_totals(const char *binary_url) return result; } -void lp_ws_cleanup() -{ - log_debug(_("cleanup CURL")); - - curl_easy_cleanup(curl); - curl_global_cleanup(); -}