mirror of
https://github.com/fangfufu/httpdirfs.git
synced 2024-09-30 14:11:16 +02:00
481 lines
14 KiB
C
481 lines
14 KiB
C
#include "network.h"
|
|
|
|
#include <ctype.h>
|
|
#include <errno.h>
|
|
#include <pthread.h>
|
|
#include <string.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
|
|
#define HTTP_OK 200
|
|
#define HTTP_PARTIAL_CONTENT 206
|
|
|
|
/* ------------------------ External variables ----------------------------*/
|
|
LinkTable *ROOT_LINK_TBL;
|
|
|
|
/* ------------------------ Static variable ------------------------------ */
|
|
/** \brief curl shared interface - not actually being used. */
|
|
static CURLSH *curl_share;
|
|
/** \brief pthread mutex for thread safety */
|
|
static pthread_mutex_t pthread_curl_lock;
|
|
/** \brief curl multi interface handle */
|
|
static CURLM *curl_multi;
|
|
|
|
/* -----------Forward declarations for static functions --------------------*/
|
|
|
|
static void HTML_to_LinkTable(GumboNode *node, LinkTable *linktbl);
|
|
static LinkType p_url_type(const char *p_url);
|
|
static Link *Link_new(const char *p_url, LinkType type);
|
|
static CURL *Link_to_curl(Link *link);
|
|
static void LinkTable_free(LinkTable *linktbl);
|
|
static void LinkTable_add(LinkTable *linktbl, Link *link);
|
|
static void LinkTable_fill(LinkTable *linktbl);
|
|
static Link *path_to_Link_recursive(char *path, LinkTable *linktbl);
|
|
static char *url_append(const char *url, const char *sublink);
|
|
static size_t WriteMemoryCallback(void *contents, size_t size, size_t nmemb,
|
|
void *userp);
|
|
static void blocking_single_transfer(CURL *curl);
|
|
static void transfer_wrapper(CURL *curl);
|
|
static int is_downloading(Link *link);
|
|
|
|
/**
|
|
* \brief convert a HTML page to a LinkTable
|
|
* \details Shamelessly copied and pasted from:
|
|
* https://github.com/google/gumbo-parser/blob/master/examples/find_links.cc
|
|
*/
|
|
static void HTML_to_LinkTable(GumboNode *node, LinkTable *linktbl);
|
|
|
|
/* -------------------------- Functions ---------------------------------- */
|
|
static int is_downloading(Link *link)
|
|
{
|
|
return link->downloading;
|
|
}
|
|
|
|
static void transfer_wrapper(CURL *curl)
|
|
{
|
|
blocking_single_transfer(curl);
|
|
}
|
|
|
|
static void pthread_lock_cb(CURL *handle, curl_lock_data data,
|
|
curl_lock_access access, void *userptr)
|
|
{
|
|
(void)access; /* unused */
|
|
(void)userptr; /* unused */
|
|
(void)handle; /* unused */
|
|
(void)data; /* unused */
|
|
pthread_mutex_lock(&pthread_curl_lock);
|
|
}
|
|
|
|
static void pthread_unlock_cb(CURL *handle, curl_lock_data data,
|
|
void *userptr)
|
|
{
|
|
(void)userptr; /* unused */
|
|
(void)handle; /* unused */
|
|
(void)data; /* unused */
|
|
pthread_mutex_unlock(&pthread_curl_lock);
|
|
}
|
|
|
|
void network_init(const char *url)
|
|
{
|
|
/* Global related */
|
|
curl_global_init(CURL_GLOBAL_ALL);
|
|
ROOT_LINK_TBL = LinkTable_new(url);
|
|
|
|
/* Share related */
|
|
curl_share = curl_share_init();
|
|
curl_share_setopt(curl_share, CURLSHOPT_SHARE, CURL_LOCK_DATA_COOKIE);
|
|
curl_share_setopt(curl_share, CURLSHOPT_SHARE, CURL_LOCK_DATA_DNS);
|
|
curl_share_setopt(curl_share, CURLSHOPT_SHARE, CURL_LOCK_DATA_CONNECT);
|
|
|
|
pthread_mutex_init(&pthread_curl_lock, NULL);
|
|
curl_share_setopt(curl_share, CURLSHOPT_LOCKFUNC, pthread_lock_cb);
|
|
curl_share_setopt(curl_share, CURLSHOPT_UNLOCKFUNC, pthread_unlock_cb);
|
|
|
|
/* Multi related */
|
|
curl_multi = curl_multi_init();
|
|
curl_multi_setopt(curl_multi, CURLMOPT_MAXCONNECTS,
|
|
CURL_MULTI_MAX_CONNECTION);
|
|
}
|
|
|
|
static CURL *Link_to_curl(Link *link)
|
|
{
|
|
#ifdef HTTPDIRFS_INFO
|
|
fprintf(stderr, "Link_to_curl(%s);\n", link->f_url);
|
|
fflush(stderr);
|
|
#endif
|
|
|
|
CURL *curl = curl_easy_init();
|
|
|
|
/* set up some basic curl stuff */
|
|
curl_easy_setopt(curl, CURLOPT_WRITEFUNCTION, WriteMemoryCallback);
|
|
curl_easy_setopt(curl, CURLOPT_USERAGENT, "mount-http-dir/libcurl");
|
|
curl_easy_setopt(curl, CURLOPT_VERBOSE, 0);
|
|
curl_easy_setopt(curl, CURLOPT_FOLLOWLOCATION, 1);
|
|
/*
|
|
* only 1 redirection is really needed
|
|
* - for following directories without the '/'
|
|
*/
|
|
curl_easy_setopt(curl, CURLOPT_MAXREDIRS, 1);
|
|
curl_easy_setopt(curl, CURLOPT_URL, link->f_url);
|
|
curl_easy_setopt(curl, CURLOPT_TCP_KEEPALIVE, 1);
|
|
curl_easy_setopt(curl, CURLOPT_SHARE, curl_share);
|
|
|
|
return curl;
|
|
}
|
|
|
|
|
|
static char *url_append(const char *url, const char *sublink)
|
|
{
|
|
int needs_separator = 0;
|
|
if (url[strlen(url)-1] != '/') {
|
|
needs_separator = 1;
|
|
}
|
|
|
|
char *str;
|
|
size_t ul = strlen(url);
|
|
size_t sl = strlen(sublink);
|
|
str = calloc(ul + sl + needs_separator + 1, sizeof(char));
|
|
strncpy(str, url, ul);
|
|
if (needs_separator) {
|
|
str[ul] = '/';
|
|
}
|
|
strncat(str, sublink, sl);
|
|
return str;
|
|
}
|
|
|
|
/* This is the single thread version */
|
|
static void blocking_single_transfer(CURL *curl)
|
|
{
|
|
#ifdef HTTPDIRFS_DEBUG
|
|
fprintf(stderr, "blocking_single_transfer(): handle %p\n", curl);
|
|
fflush(stderr);
|
|
#endif
|
|
CURLcode res = curl_easy_perform(curl);
|
|
if (res) {
|
|
fprintf(stderr, "blocking_single_transfer() failed: %u, %s\n", res,
|
|
curl_easy_strerror(res));
|
|
fflush(stderr);
|
|
}
|
|
#ifdef HTTPDIRFS_DEBUG
|
|
fprintf(stderr, "blocking_single_transfer(): DONE\n");
|
|
fflush(stderr);
|
|
#endif
|
|
}
|
|
|
|
static size_t
|
|
WriteMemoryCallback(void *contents, size_t size, size_t nmemb, void *userp)
|
|
{
|
|
size_t realsize = size * nmemb;
|
|
struct MemoryStruct *mem = (struct MemoryStruct *)userp;
|
|
|
|
mem->memory = realloc(mem->memory, mem->size + realsize + 1);
|
|
if(mem->memory == NULL) {
|
|
/* out of memory! */
|
|
fprintf(stderr, "WriteMemoryCallback(): cannot realloc!\n");
|
|
fflush(stderr);
|
|
return 0;
|
|
}
|
|
|
|
memmove(&(mem->memory[mem->size]), contents, realsize);
|
|
mem->size += realsize;
|
|
mem->memory[mem->size] = 0;
|
|
|
|
return realsize;
|
|
}
|
|
|
|
static Link *Link_new(const char *p_url, LinkType type)
|
|
{
|
|
Link *link = calloc(1, sizeof(Link));
|
|
strncpy(link->p_url, p_url, LINK_LEN_MAX);
|
|
link->type = type;
|
|
|
|
/* remove the '/' from p_url if it exists */
|
|
char *c = &(link->p_url[strnlen(link->p_url, LINK_LEN_MAX) - 1]);
|
|
if ( *c == '/') {
|
|
*c = '\0';
|
|
}
|
|
|
|
return link;
|
|
}
|
|
|
|
long Link_download(const char *path, char *output_buf, size_t size,
|
|
off_t offset)
|
|
{
|
|
Link *link;
|
|
link = path_to_Link(path);
|
|
if (!link) {
|
|
return -ENOENT;
|
|
}
|
|
|
|
size_t start = offset;
|
|
size_t end = start + size;
|
|
CURL *curl = Link_to_curl(link);
|
|
char range_str[64];
|
|
snprintf(range_str, sizeof(range_str), "%lu-%lu", start, end);
|
|
|
|
MemoryStruct buf;
|
|
buf.size = 0;
|
|
buf.memory = NULL;
|
|
curl_easy_setopt(curl, CURLOPT_WRITEDATA, (void *)&buf);
|
|
curl_easy_setopt(curl, CURLOPT_RANGE, range_str);
|
|
|
|
transfer_wrapper(curl);
|
|
|
|
long http_resp;
|
|
curl_easy_getinfo(curl, CURLINFO_RESPONSE_CODE, &http_resp);
|
|
if ( (http_resp != HTTP_OK) && ( http_resp != HTTP_PARTIAL_CONTENT) ) {
|
|
fprintf(stderr, "Link_download(): Could not download %s, HTTP %ld\n",
|
|
link->f_url, http_resp);
|
|
fflush(stderr);
|
|
return -ENOENT;
|
|
}
|
|
|
|
double dl;
|
|
curl_easy_getinfo(curl, CURLINFO_SIZE_DOWNLOAD, &dl);
|
|
|
|
size_t recv = dl;
|
|
if (recv > size) {
|
|
recv = size;
|
|
}
|
|
|
|
memmove(output_buf, buf.memory, recv);
|
|
curl_easy_cleanup(curl);
|
|
free(buf.memory);
|
|
return recv;
|
|
}
|
|
|
|
LinkTable *LinkTable_new(const char *url)
|
|
{
|
|
LinkTable *linktbl = calloc(1, sizeof(LinkTable));
|
|
|
|
/* populate the base URL */
|
|
LinkTable_add(linktbl, Link_new("/", LINK_HEAD));
|
|
Link *head_link = linktbl->links[0];
|
|
head_link->type = LINK_HEAD;
|
|
strncpy(head_link->f_url, url, URL_LEN_MAX);
|
|
|
|
/* start downloading the base URL */
|
|
CURL *curl = Link_to_curl(head_link);
|
|
MemoryStruct buf;
|
|
buf.size = 0;
|
|
buf.memory = NULL;
|
|
curl_easy_setopt(curl, CURLOPT_WRITEDATA, (void *)&buf);
|
|
|
|
transfer_wrapper(curl);
|
|
|
|
/* if downloading base URL failed */
|
|
long http_resp;
|
|
curl_easy_getinfo(curl, CURLINFO_RESPONSE_CODE, &http_resp);
|
|
if (http_resp != HTTP_OK) {
|
|
fprintf(stderr, "link.c: LinkTable_new() cannot retrive the base URL, \
|
|
URL: %s, HTTP %ld\n", url, http_resp);
|
|
fflush(stderr);
|
|
LinkTable_free(linktbl);
|
|
linktbl = NULL;
|
|
return linktbl;
|
|
};
|
|
curl_easy_cleanup(curl);
|
|
|
|
/* Otherwise parsed the received data */
|
|
GumboOutput* output = gumbo_parse(buf.memory);
|
|
HTML_to_LinkTable(output->root, linktbl);
|
|
gumbo_destroy_output(&kGumboDefaultOptions, output);
|
|
free(buf.memory);
|
|
|
|
/* Fill in the link table */
|
|
LinkTable_fill(linktbl);
|
|
#ifdef HTTPDIRFS_INFO
|
|
LinkTable_print(linktbl);
|
|
#endif
|
|
return linktbl;
|
|
}
|
|
|
|
static void LinkTable_free(LinkTable *linktbl)
|
|
{
|
|
for (int i = 0; i < linktbl->num; i++) {
|
|
free(linktbl->links[i]);
|
|
}
|
|
free(linktbl->links);
|
|
free(linktbl);
|
|
}
|
|
|
|
static void LinkTable_add(LinkTable *linktbl, Link *link)
|
|
{
|
|
linktbl->num++;
|
|
linktbl->links = realloc(
|
|
linktbl->links,
|
|
linktbl->num * sizeof(Link *));
|
|
linktbl->links[linktbl->num - 1] = link;
|
|
}
|
|
|
|
size_t Link_get_size(Link *this_link)
|
|
{
|
|
double cl = 0;
|
|
|
|
if (this_link->type == LINK_FILE) {
|
|
CURL *curl = Link_to_curl(this_link);
|
|
curl_easy_setopt(curl, CURLOPT_NOBODY, 1);
|
|
transfer_wrapper(curl);
|
|
long http_resp;
|
|
curl_easy_getinfo(curl, CURLINFO_RESPONSE_CODE, &http_resp);
|
|
if (http_resp == HTTP_OK) {
|
|
curl_easy_getinfo(curl, CURLINFO_CONTENT_LENGTH_DOWNLOAD, &cl);
|
|
if (cl == -1) {
|
|
/* Turns out not to be a file after all */
|
|
this_link->content_length = 0;
|
|
this_link->type = LINK_DIR;
|
|
} else {
|
|
this_link->content_length = cl;
|
|
this_link->type = LINK_FILE;
|
|
}
|
|
} else {
|
|
this_link->type = LINK_INVALID;
|
|
}
|
|
curl_easy_cleanup(curl);
|
|
}
|
|
|
|
return cl;
|
|
}
|
|
|
|
void LinkTable_fill(LinkTable *linktbl)
|
|
{
|
|
Link *head_link = linktbl->links[0];
|
|
for (int i = 0; i < linktbl->num; i++) {
|
|
Link *this_link = linktbl->links[i];
|
|
if (this_link->type) {
|
|
char *url;
|
|
url = url_append(head_link->f_url, this_link->p_url);
|
|
strncpy(this_link->f_url, url, URL_LEN_MAX);
|
|
free(url);
|
|
if (this_link->type == LINK_FILE) {
|
|
this_link->content_length = Link_get_size(this_link);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
#ifdef HTTPDIRFS_INFO
|
|
void LinkTable_print(LinkTable *linktbl)
|
|
{
|
|
fprintf(stderr, "--------------------------------------------\n");
|
|
fprintf(stderr, " LinkTable %p\n", linktbl);
|
|
fprintf(stderr, "--------------------------------------------\n");
|
|
for (int i = 0; i < linktbl->num; i++) {
|
|
Link *this_link = linktbl->links[i];
|
|
fprintf(stderr, "%d %c %lu %s %s\n",
|
|
i,
|
|
this_link->type,
|
|
this_link->content_length,
|
|
this_link->p_url,
|
|
this_link->f_url
|
|
);
|
|
fflush(stderr);
|
|
}
|
|
fprintf(stderr, "--------------------------------------------\n");
|
|
}
|
|
#endif
|
|
|
|
static LinkType p_url_type(const char *p_url)
|
|
{
|
|
/* The link name has to start with alphanumerical character */
|
|
if (!isalnum(p_url[0])) {
|
|
return LINK_INVALID;
|
|
}
|
|
|
|
/* check for http:// and https:// */
|
|
if ( !strncmp(p_url, "http://", 7) || !strncmp(p_url, "https://", 8) ) {
|
|
return LINK_INVALID;
|
|
}
|
|
|
|
if ( p_url[strlen(p_url) - 1] == '/' ) {
|
|
return LINK_DIR;
|
|
}
|
|
|
|
return LINK_FILE;
|
|
}
|
|
|
|
static void HTML_to_LinkTable(GumboNode *node, LinkTable *linktbl)
|
|
{
|
|
if (node->type != GUMBO_NODE_ELEMENT) {
|
|
return;
|
|
}
|
|
GumboAttribute* href;
|
|
|
|
if (node->v.element.tag == GUMBO_TAG_A &&
|
|
(href = gumbo_get_attribute(&node->v.element.attributes, "href"))) {
|
|
/* if it is valid, copy the link onto the heap */
|
|
LinkType type = p_url_type(href->value);
|
|
if (type) {
|
|
LinkTable_add(linktbl, Link_new(href->value, type));
|
|
}
|
|
}
|
|
|
|
/* Note the recursive call, lol. */
|
|
GumboVector *children = &node->v.element.children;
|
|
for (size_t i = 0; i < children->length; ++i) {
|
|
HTML_to_LinkTable((GumboNode*)children->data[i], linktbl);
|
|
}
|
|
return;
|
|
}
|
|
|
|
static Link *path_to_Link_recursive(char *path, LinkTable *linktbl)
|
|
{
|
|
/* skip the leading '/' if it exists */
|
|
if (*path == '/') {
|
|
path++;
|
|
}
|
|
|
|
/* remove the last '/' if it exists */
|
|
char *slash = &(path[strnlen(path, URL_LEN_MAX) - 1]);
|
|
if (*slash == '/') {
|
|
*slash = '\0';
|
|
}
|
|
|
|
slash = strchr(path, '/');
|
|
if ( slash == NULL ) {
|
|
/* We cannot find another '/', we have reached the last level */
|
|
for (int i = 1; i < linktbl->num; i++) {
|
|
if (!strncmp(path, linktbl->links[i]->p_url, LINK_LEN_MAX)) {
|
|
/* We found our link */
|
|
return linktbl->links[i];
|
|
}
|
|
}
|
|
} else {
|
|
/*
|
|
* We can still find '/', time to consume the path and traverse
|
|
* the tree structure
|
|
*/
|
|
|
|
/*
|
|
* add termination mark to the current string,
|
|
* effective create two substrings
|
|
*/
|
|
*slash = '\0';
|
|
/* move the pointer past the '/' */
|
|
char *next_path = slash + 1;
|
|
for (int i = 1; i < linktbl->num; i++) {
|
|
if (!strncmp(path, linktbl->links[i]->p_url, LINK_LEN_MAX)) {
|
|
/* The next sub-directory exists */
|
|
LinkTable *next_table = linktbl->links[i]->next_table;
|
|
if (!(next_table)) {
|
|
next_table = LinkTable_new(linktbl->links[i]->f_url);
|
|
}
|
|
return path_to_Link_recursive(next_path, next_table);
|
|
}
|
|
}
|
|
}
|
|
#ifdef HTTPDIRFS_DEBUG
|
|
fprintf(stderr, "path_to_Link(): %s does not exist.\n", path);
|
|
fflush(stderr);
|
|
#endif
|
|
return NULL;
|
|
}
|
|
|
|
Link *path_to_Link(const char *path)
|
|
{
|
|
char *new_path = strndup(path, URL_LEN_MAX);
|
|
return path_to_Link_recursive(new_path, ROOT_LINK_TBL);
|
|
}
|
|
|
|
|