X-Git-Url: http://git.wpitchoune.net/gitweb/?a=blobdiff_plain;f=src%2Flp_ws.c;h=fcf08c41b3ce736e1a66510b510bada33df04b8f;hb=cbe5f873c5f0306a7a5d2f431992dbbb6edbe332;hp=4549d92c8ab4b2855cae533459e9b5062ca08198;hpb=fa460465ab78c48ace6d8563e6d35917b30fab09;p=ppastats.git diff --git a/src/lp_ws.c b/src/lp_ws.c index 4549d92..fcf08c4 100644 --- a/src/lp_ws.c +++ b/src/lp_ws.c @@ -20,171 +20,150 @@ #include #define _(String) gettext(String) +#include #include #include -#include -#include #include #include "cache.h" +#include "fcache.h" +#include "http.h" #include "list.h" #include "log.h" #include "lp_ws.h" #include "lp_json.h" #include "ppastats.h" -static const char *QUERY_GET_PUBLISHED_BINARIES = "?ws.op=getPublishedBinaries"; +static const char * +QUERY_GET_PUBLISHED_BINARIES = "?ws.op=getPublishedBinaries&ws.size=150"; static const char *QUERY_GET_DOWNLOAD_COUNT = "?ws.op=getDownloadCount"; static const char * QUERY_GET_DAILY_DOWNLOAD_TOTALS = "?ws.op=getDailyDownloadTotals"; -static const int DEFAULT_FETCH_RETRIES = 3; - -static CURL *curl; +static json_object *get_json_object(const char *url) +{ + json_object *obj = NULL; + char *body; -struct ucontent { - char *data; - size_t len; -}; + body = get_url_content(url, 0); -static size_t cbk_curl(void *buffer, size_t size, size_t nmemb, void *userp) -{ - size_t realsize = size * nmemb; - struct ucontent *mem = (struct ucontent *)userp; + if (body) { + obj = json_tokener_parse(body); - mem->data = realloc(mem->data, mem->len + realsize + 1); + free(body); - memcpy(&(mem->data[mem->len]), buffer, realsize); - mem->len += realsize; - mem->data[mem->len] = 0; + return obj; + } - return realsize; + return NULL; } -static char *fetch_url(const char *url) +char *get_bpph_list_cache_key(const char *archive_url) { - struct ucontent *content = malloc(sizeof(struct ucontent)); - char *result; - long code; - int retries; + char *key; - log_debug(_("fetch_url(): %s"), url); + key = malloc(strlen(archive_url + 7) + strlen("/bpph") + 1); + sprintf(key, "%s/bpph", archive_url + 7); - if (!curl) { - log_debug(_("initializing CURL")); - curl_global_init(CURL_GLOBAL_ALL); - curl = curl_easy_init(); - } + return key; +} - if (!curl) - exit(EXIT_FAILURE); +struct bpph **get_bpph_list_from_cache(const char *key) +{ + char *content; + struct bpph **list; + json_object *json; - result = NULL; + content = fcache_get(key); + if (!content) + return NULL; - retries = DEFAULT_FETCH_RETRIES; + json = json_tokener_parse(content); + if (!json) + return NULL; - retrieve: - content->data = malloc(1); - content->data[0] = '\0'; - content->len = 0; + list = json_object_to_bpph_list(json); - curl_easy_setopt(curl, CURLOPT_URL, url); - curl_easy_setopt(curl, CURLOPT_VERBOSE, 0); - curl_easy_setopt(curl, CURLOPT_WRITEFUNCTION, cbk_curl); - curl_easy_setopt(curl, CURLOPT_WRITEDATA, content); - curl_easy_setopt(curl, CURLOPT_USERAGENT, "ppastats/0.0"); + json_object_put(json); + free(content); - if (curl_easy_perform(curl) == CURLE_OK) { - curl_easy_getinfo(curl, CURLINFO_RESPONSE_CODE, &code); + return list; +} - switch (code) { - case 200: - result = content->data; - break; - case 500: - case 502: - case 503: - case 504: - if (retries) { - log_err(_("Fetch failed with code %ld " - "for URL %s"), - code, - url); - - log_debug(_("Wait 5s before retry")); - sleep(5); - - free(content->data); - retries--; - goto retrieve; - } - default: - log_err(_("Fetch failed: %ld"), code); - } - } +struct tm *get_last_creation_date(struct bpph **list) +{ + time_t last, t; + struct bpph **cur; - if (!result) - free(content->data); + last = 0; - free(content); + if (list) + for (cur = list; *cur; cur++) { + t = mktime(&(*cur)->date_created); + if (t > last) + last = t; + } - return result; + return localtime(&last); } -static json_object *get_json_object(const char *url) +struct bpph **get_bpph_list(const char *archive_url, const char *pkg_status) { - json_object *obj = NULL; - char *body; - - body = fetch_url(url); - - if (body) { - obj = json_tokener_parse(body); + char *url, *key, *tmp; + struct bpph **result = NULL; + struct json_object *o, *bpph_json, *o_next; + char *created_since_date; + struct tm *tm; + int ok; + + url = malloc(strlen(archive_url) + + strlen(QUERY_GET_PUBLISHED_BINARIES) + + 1); + strcpy(url, archive_url); + strcat(url, QUERY_GET_PUBLISHED_BINARIES); - free(body); + key = get_bpph_list_cache_key(archive_url); - return obj; - } + result = get_bpph_list_from_cache(key); - return NULL; -} + if (result) { + tm = get_last_creation_date(result); -#define json_object_to_bpph_list \ -json_object_to_binary_package_publishing_history_list + created_since_date = malloc(200); + strftime(created_since_date, + 100, + "%FT%T", + tm); -struct binary_package_publishing_history * * -get_binary_package_publishing_history_list(const char *archive_url, - const char *pkg_status) -{ - struct json_object *o_next; - char *url; - json_object *o; - void **result = NULL; + printf("Update package since: %s\n", created_since_date); - url = malloc(strlen(archive_url)+ - strlen(QUERY_GET_PUBLISHED_BINARIES)+ - (pkg_status ? strlen("&status=")+strlen(pkg_status) : 0)+ - 1); + tmp = malloc(strlen(url) + + strlen("&created_since_date=") + + strlen(created_since_date)+1); + strcpy(tmp, url); + strcat(tmp, "&created_since_date="); + strcat(tmp, created_since_date); - strcpy(url, archive_url); - strcat(url, QUERY_GET_PUBLISHED_BINARIES); + free(url); + url = tmp; - if (pkg_status) { - strcat(url, "&status="); - strcat(url, pkg_status); + free(created_since_date); } + ok = 1; while (url) { o = get_json_object(url); free(url); url = NULL; - if (!o) + if (!o) { + ok = 0; break; + } - result = list_append_list(result, - (void **)json_object_to_bpph_list(o)); + result = bpph_list_append_list(result, + json_object_to_bpph_list(o)); o_next = json_object_object_get(o, "next_collection_link"); @@ -192,9 +171,18 @@ get_binary_package_publishing_history_list(const char *archive_url, url = strdup(json_object_get_string(o_next)); json_object_put(o); + + } + + if (ok) { + bpph_json = bpph_list_to_json(result); + fcache_put(key, json_object_to_json_string(bpph_json)); + json_object_put(bpph_json); } - return (struct binary_package_publishing_history **)result; + free(key); + + return result; } int get_download_count(const char *archive_url) @@ -224,12 +212,20 @@ const struct distro_arch_series *get_distro_arch_series(const char *url) { json_object *obj; const struct distro_arch_series *distro; + char *content; distro = cache_get(url); if (distro) return (struct distro_arch_series *)distro; - obj = get_json_object(url); + content = get_url_content(url, 1); + + if (!content) + return NULL; + + obj = json_tokener_parse(content); + + free(content); if (!obj) return NULL; @@ -247,12 +243,20 @@ const struct distro_series *get_distro_series(const char *url) { json_object *obj; const struct distro_series *distro; + char *content; distro = cache_get(url); if (distro) return (struct distro_series *)distro; - obj = get_json_object(url); + content = get_url_content(url, 1); + + if (!content) + return NULL; + + obj = json_tokener_parse(content); + + free(content); if (!obj) return NULL; @@ -290,10 +294,3 @@ struct daily_download_total **get_daily_download_totals(const char *binary_url) return result; } -void lp_ws_cleanup() -{ - log_debug(_("cleanup CURL")); - - curl_easy_cleanup(curl); - curl_global_cleanup(); -}