mirror of
https://codeberg.org/grunfink/snac2.git
synced 2024-12-27 01:33:36 +00:00
833 lines
22 KiB
C
833 lines
22 KiB
C
/* snac - A simple, minimalistic ActivityPub instance */
|
|
/* copyright (c) 2022 - 2024 grunfink et al. / MIT license */
|
|
|
|
#include "xs.h"
|
|
#include "xs_io.h"
|
|
#include "xs_json.h"
|
|
#include "xs_socket.h"
|
|
#include "xs_httpd.h"
|
|
#include "xs_mime.h"
|
|
#include "xs_time.h"
|
|
#include "xs_openssl.h"
|
|
#include "xs_fcgi.h"
|
|
#include "xs_html.h"
|
|
|
|
#include "snac.h"
|
|
|
|
#include <setjmp.h>
|
|
#include <pthread.h>
|
|
#include <semaphore.h>
|
|
#include <fcntl.h>
|
|
#include <stdint.h>
|
|
|
|
#include <sys/resource.h> // for getrlimit()
|
|
|
|
#include <sys/mman.h>
|
|
|
|
#ifdef USE_POLL_FOR_SLEEP
|
|
#include <poll.h>
|
|
#endif
|
|
|
|
/** server state **/
|
|
srv_state *p_state = NULL;
|
|
|
|
|
|
/** job control **/
|
|
|
|
/* mutex to access the lists of jobs */
|
|
static pthread_mutex_t job_mutex;
|
|
|
|
/* semaphore to trigger job processing */
|
|
static sem_t *job_sem;
|
|
|
|
typedef struct job_fifo_item {
|
|
struct job_fifo_item *next;
|
|
xs_val *job;
|
|
} job_fifo_item;
|
|
|
|
static job_fifo_item *job_fifo_first = NULL;
|
|
static job_fifo_item *job_fifo_last = NULL;
|
|
|
|
|
|
/** other global data **/
|
|
|
|
static jmp_buf on_break;
|
|
|
|
|
|
/** code **/
|
|
|
|
/* nodeinfo 2.0 template */
|
|
const char *nodeinfo_2_0_template = ""
|
|
"{\"version\":\"2.0\","
|
|
"\"software\":{\"name\":\"snac\",\"version\":\"" VERSION "\"},"
|
|
"\"protocols\":[\"activitypub\"],"
|
|
"\"services\":{\"outbound\":[],\"inbound\":[]},"
|
|
"\"usage\":{\"users\":{\"total\":%d,\"activeMonth\":%d,\"activeHalfyear\":%d},"
|
|
"\"localPosts\":%d},"
|
|
"\"openRegistrations\":false,\"metadata\":{}}";
|
|
|
|
xs_str *nodeinfo_2_0(void)
|
|
/* builds a nodeinfo json object */
|
|
{
|
|
int n_utotal = 0;
|
|
int n_umonth = 0;
|
|
int n_uhyear = 0;
|
|
int n_posts = 0;
|
|
xs *users = user_list();
|
|
xs_list *p = users;
|
|
char *v;
|
|
double now = (double)time(NULL);
|
|
|
|
while (xs_list_iter(&p, &v)) {
|
|
/* build the full path name to the last usage log */
|
|
xs *llfn = xs_fmt("%s/user/%s/lastlog.txt", srv_basedir, v);
|
|
double llsecs = now - mtime(llfn);
|
|
|
|
if (llsecs < 60 * 60 * 24 * 30 * 6) {
|
|
n_uhyear++;
|
|
|
|
if (llsecs < 60 * 60 * 24 * 30)
|
|
n_umonth++;
|
|
}
|
|
|
|
n_utotal++;
|
|
|
|
/* build the file to each user public.idx */
|
|
xs *pidxfn = xs_fmt("%s/user/%s/public.idx", srv_basedir, v);
|
|
n_posts += index_len(pidxfn);
|
|
}
|
|
|
|
return xs_fmt(nodeinfo_2_0_template, n_utotal, n_umonth, n_uhyear, n_posts);
|
|
}
|
|
|
|
|
|
static xs_str *greeting_html(void)
|
|
/* processes and returns greeting.html */
|
|
{
|
|
/* try to open greeting.html */
|
|
xs *fn = xs_fmt("%s/greeting.html", srv_basedir);
|
|
FILE *f;
|
|
xs_str *s = NULL;
|
|
|
|
if ((f = fopen(fn, "r")) != NULL) {
|
|
s = xs_readall(f);
|
|
fclose(f);
|
|
|
|
/* replace %host% */
|
|
s = xs_replace_i(s, "%host%", xs_dict_get(srv_config, "host"));
|
|
|
|
const char *adm_email = xs_dict_get(srv_config, "admin_email");
|
|
if (xs_is_null(adm_email) || *adm_email == '\0')
|
|
adm_email = "the administrator of this instance";
|
|
|
|
/* replace %admin_email */
|
|
s = xs_replace_i(s, "%admin_email%", adm_email);
|
|
|
|
/* does it have a %userlist% mark? */
|
|
if (xs_str_in(s, "%userlist%") != -1) {
|
|
char *host = xs_dict_get(srv_config, "host");
|
|
xs *list = user_list();
|
|
xs_list *p = list;
|
|
xs_str *uid;
|
|
|
|
xs_html *ul = xs_html_tag("ul",
|
|
xs_html_attr("class", "snac-user-list"));
|
|
|
|
p = list;
|
|
while (xs_list_iter(&p, &uid)) {
|
|
snac user;
|
|
|
|
if (user_open(&user, uid)) {
|
|
xs_html_add(ul,
|
|
xs_html_tag("li",
|
|
xs_html_tag("a",
|
|
xs_html_attr("href", user.actor),
|
|
xs_html_text("@"),
|
|
xs_html_text(uid),
|
|
xs_html_text("@"),
|
|
xs_html_text(host),
|
|
xs_html_text(" ("),
|
|
xs_html_text(xs_dict_get(user.config, "name")),
|
|
xs_html_text(")"))));
|
|
|
|
user_free(&user);
|
|
}
|
|
}
|
|
|
|
xs *s1 = xs_html_render(ul);
|
|
s = xs_replace_i(s, "%userlist%", s1);
|
|
}
|
|
}
|
|
|
|
return s;
|
|
}
|
|
|
|
|
|
int server_get_handler(xs_dict *req, const char *q_path,
|
|
char **body, int *b_size, char **ctype)
|
|
/* basic server services */
|
|
{
|
|
int status = 0;
|
|
|
|
(void)req;
|
|
|
|
/* is it the server root? */
|
|
if (*q_path == '\0') {
|
|
xs_dict *q_vars = xs_dict_get(req, "q_vars");
|
|
char *t = NULL;
|
|
|
|
if (xs_type(q_vars) == XSTYPE_DICT && (t = xs_dict_get(q_vars, "t"))) {
|
|
/** search by tag **/
|
|
int skip = 0;
|
|
int show = xs_number_get(xs_dict_get(srv_config, "max_timeline_entries"));
|
|
char *v;
|
|
|
|
if ((v = xs_dict_get(q_vars, "skip")) != NULL)
|
|
skip = atoi(v);
|
|
if ((v = xs_dict_get(q_vars, "show")) != NULL)
|
|
show = atoi(v);
|
|
|
|
xs *tl = tag_search(t, skip, show + 1);
|
|
int more = 0;
|
|
if (xs_list_len(tl) >= show + 1) {
|
|
/* drop the last one */
|
|
tl = xs_list_del(tl, -1);
|
|
more = 1;
|
|
}
|
|
|
|
*body = html_timeline(NULL, tl, 0, skip, show, more, t, NULL, 0);
|
|
}
|
|
else
|
|
if (xs_type(xs_dict_get(srv_config, "show_instance_timeline")) == XSTYPE_TRUE) {
|
|
/** instance timeline **/
|
|
xs *tl = timeline_instance_list(0, 30);
|
|
*body = html_timeline(NULL, tl, 0, 0, 0, 0, NULL, NULL, 0);
|
|
}
|
|
else
|
|
*body = greeting_html();
|
|
|
|
if (*body)
|
|
status = 200;
|
|
}
|
|
else
|
|
if (strcmp(q_path, "/susie.png") == 0 || strcmp(q_path, "/favicon.ico") == 0 ) {
|
|
status = 200;
|
|
*body = xs_base64_dec(default_avatar_base64(), b_size);
|
|
*ctype = "image/png";
|
|
}
|
|
else
|
|
if (strcmp(q_path, "/.well-known/nodeinfo") == 0) {
|
|
status = 200;
|
|
*ctype = "application/json; charset=utf-8";
|
|
*body = xs_fmt("{\"links\":["
|
|
"{\"rel\":\"http:/" "/nodeinfo.diaspora.software/ns/schema/2.0\","
|
|
"\"href\":\"%s/nodeinfo_2_0\"}]}",
|
|
srv_baseurl);
|
|
}
|
|
else
|
|
if (strcmp(q_path, "/.well-known/host-meta") == 0) {
|
|
status = 200;
|
|
*ctype = "application/xrd+xml";
|
|
*body = xs_fmt("<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n"
|
|
"<XRD>"
|
|
"<Link rel=\"lrdd\" type=\"application/xrd+xml\" template=\"https://%s/.well-known/webfinger?resource={uri}\"/>"
|
|
"</XRD>", xs_dict_get(srv_config, "host"));
|
|
}
|
|
else
|
|
if (strcmp(q_path, "/nodeinfo_2_0") == 0) {
|
|
status = 200;
|
|
*ctype = "application/json; charset=utf-8";
|
|
*body = nodeinfo_2_0();
|
|
}
|
|
else
|
|
if (strcmp(q_path, "/robots.txt") == 0) {
|
|
status = 200;
|
|
*ctype = "text/plain";
|
|
*body = xs_str_new("User-agent: *\n"
|
|
"Disallow: /\n");
|
|
}
|
|
|
|
if (status != 0)
|
|
srv_debug(1, xs_fmt("server_get_handler serving '%s' %d", q_path, status));
|
|
|
|
return status;
|
|
}
|
|
|
|
|
|
void httpd_connection(FILE *f)
|
|
/* the connection processor */
|
|
{
|
|
xs *req;
|
|
char *method;
|
|
int status = 0;
|
|
xs_str *body = NULL;
|
|
int b_size = 0;
|
|
char *ctype = NULL;
|
|
xs *headers = xs_dict_new();
|
|
xs *q_path = NULL;
|
|
xs *payload = NULL;
|
|
xs *etag = NULL;
|
|
int p_size = 0;
|
|
char *p;
|
|
int fcgi_id;
|
|
|
|
if (p_state->use_fcgi)
|
|
req = xs_fcgi_request(f, &payload, &p_size, &fcgi_id);
|
|
else
|
|
req = xs_httpd_request(f, &payload, &p_size);
|
|
|
|
if (req == NULL) {
|
|
/* probably because a timeout */
|
|
fclose(f);
|
|
return;
|
|
}
|
|
|
|
if (!(method = xs_dict_get(req, "method")) || !(p = xs_dict_get(req, "path"))) {
|
|
/* missing needed headers; discard */
|
|
fclose(f);
|
|
return;
|
|
}
|
|
|
|
q_path = xs_dup(p);
|
|
|
|
/* crop the q_path from leading / and the prefix */
|
|
if (xs_endswith(q_path, "/"))
|
|
q_path = xs_crop_i(q_path, 0, -1);
|
|
|
|
p = xs_dict_get(srv_config, "prefix");
|
|
if (xs_startswith(q_path, p))
|
|
q_path = xs_crop_i(q_path, strlen(p), 0);
|
|
|
|
if (strcmp(method, "GET") == 0 || strcmp(method, "HEAD") == 0) {
|
|
/* cascade through */
|
|
if (status == 0)
|
|
status = server_get_handler(req, q_path, &body, &b_size, &ctype);
|
|
|
|
if (status == 0)
|
|
status = webfinger_get_handler(req, q_path, &body, &b_size, &ctype);
|
|
|
|
if (status == 0)
|
|
status = activitypub_get_handler(req, q_path, &body, &b_size, &ctype);
|
|
|
|
#ifndef NO_MASTODON_API
|
|
if (status == 0)
|
|
status = oauth_get_handler(req, q_path, &body, &b_size, &ctype);
|
|
|
|
if (status == 0)
|
|
status = mastoapi_get_handler(req, q_path, &body, &b_size, &ctype);
|
|
#endif /* NO_MASTODON_API */
|
|
|
|
if (status == 0)
|
|
status = html_get_handler(req, q_path, &body, &b_size, &ctype, &etag);
|
|
}
|
|
else
|
|
if (strcmp(method, "POST") == 0) {
|
|
|
|
#ifndef NO_MASTODON_API
|
|
if (status == 0)
|
|
status = oauth_post_handler(req, q_path,
|
|
payload, p_size, &body, &b_size, &ctype);
|
|
|
|
if (status == 0)
|
|
status = mastoapi_post_handler(req, q_path,
|
|
payload, p_size, &body, &b_size, &ctype);
|
|
#endif
|
|
|
|
if (status == 0)
|
|
status = activitypub_post_handler(req, q_path,
|
|
payload, p_size, &body, &b_size, &ctype);
|
|
|
|
if (status == 0)
|
|
status = html_post_handler(req, q_path,
|
|
payload, p_size, &body, &b_size, &ctype);
|
|
}
|
|
else
|
|
if (strcmp(method, "PUT") == 0) {
|
|
|
|
#ifndef NO_MASTODON_API
|
|
if (status == 0)
|
|
status = mastoapi_put_handler(req, q_path,
|
|
payload, p_size, &body, &b_size, &ctype);
|
|
#endif
|
|
|
|
}
|
|
else
|
|
if (strcmp(method, "OPTIONS") == 0) {
|
|
status = 200;
|
|
}
|
|
else
|
|
if (strcmp(method, "DELETE") == 0) {
|
|
#ifndef NO_MASTODON_API
|
|
if (status == 0)
|
|
status = mastoapi_delete_handler(req, q_path,
|
|
&body, &b_size, &ctype);
|
|
#endif
|
|
}
|
|
|
|
/* unattended? it's an error */
|
|
if (status == 0) {
|
|
srv_archive_error("unattended_method", "unattended method", req, payload);
|
|
srv_debug(1, xs_fmt("httpd_connection unattended %s %s", method, q_path));
|
|
status = 404;
|
|
}
|
|
|
|
if (status == 403)
|
|
body = xs_str_new("<h1>403 Forbidden</h1>");
|
|
|
|
if (status == 404)
|
|
body = xs_str_new("<h1>404 Not Found</h1>");
|
|
|
|
if (status == 400 && body != NULL)
|
|
body = xs_str_new("<h1>400 Bad Request</h1>");
|
|
|
|
if (status == 303)
|
|
headers = xs_dict_append(headers, "location", body);
|
|
|
|
if (status == 401) {
|
|
xs *www_auth = xs_fmt("Basic realm=\"@%s@%s snac login\"",
|
|
body, xs_dict_get(srv_config, "host"));
|
|
|
|
headers = xs_dict_append(headers, "WWW-Authenticate", www_auth);
|
|
}
|
|
|
|
if (ctype == NULL)
|
|
ctype = "text/html; charset=utf-8";
|
|
|
|
headers = xs_dict_append(headers, "content-type", ctype);
|
|
headers = xs_dict_append(headers, "x-creator", USER_AGENT);
|
|
|
|
if (!xs_is_null(etag))
|
|
headers = xs_dict_append(headers, "etag", etag);
|
|
|
|
/* if there are any additional headers, add them */
|
|
xs_dict *more_headers = xs_dict_get(srv_config, "http_headers");
|
|
if (xs_type(more_headers) == XSTYPE_DICT) {
|
|
char *k, *v;
|
|
int c = 0;
|
|
while (xs_dict_next(more_headers, &k, &v, &c))
|
|
headers = xs_dict_set(headers, k, v);
|
|
}
|
|
|
|
if (b_size == 0 && body != NULL)
|
|
b_size = strlen(body);
|
|
|
|
/* if it was a HEAD, no body will be sent */
|
|
if (strcmp(method, "HEAD") == 0)
|
|
body = xs_free(body);
|
|
|
|
headers = xs_dict_append(headers, "access-control-allow-origin", "*");
|
|
headers = xs_dict_append(headers, "access-control-allow-headers", "*");
|
|
|
|
if (p_state->use_fcgi)
|
|
xs_fcgi_response(f, status, headers, body, b_size, fcgi_id);
|
|
else
|
|
xs_httpd_response(f, status, headers, body, b_size);
|
|
|
|
fclose(f);
|
|
|
|
srv_archive("RECV", NULL, req, payload, p_size, status, headers, body, b_size);
|
|
|
|
/* JSON validation check */
|
|
if (!xs_is_null(body) && strcmp(ctype, "application/json") == 0) {
|
|
xs *j = xs_json_loads(body);
|
|
|
|
if (j == NULL) {
|
|
srv_log(xs_fmt("bad JSON"));
|
|
srv_archive_error("bad_json", "bad JSON", req, body);
|
|
}
|
|
}
|
|
|
|
xs_free(body);
|
|
}
|
|
|
|
|
|
void job_post(const xs_val *job, int urgent)
|
|
/* posts a job for the threads to process it */
|
|
{
|
|
if (job != NULL) {
|
|
/* lock the mutex */
|
|
pthread_mutex_lock(&job_mutex);
|
|
|
|
job_fifo_item *i = xs_realloc(NULL, sizeof(job_fifo_item));
|
|
*i = (job_fifo_item){ NULL, xs_dup(job) };
|
|
|
|
if (job_fifo_first == NULL)
|
|
job_fifo_first = job_fifo_last = i;
|
|
else
|
|
if (urgent) {
|
|
/* prepend */
|
|
i->next = job_fifo_first;
|
|
job_fifo_first = i;
|
|
}
|
|
else {
|
|
/* append */
|
|
job_fifo_last->next = i;
|
|
job_fifo_last = i;
|
|
}
|
|
|
|
p_state->job_fifo_size++;
|
|
|
|
if (p_state->job_fifo_size > p_state->peak_job_fifo_size)
|
|
p_state->peak_job_fifo_size = p_state->job_fifo_size;
|
|
|
|
/* unlock the mutex */
|
|
pthread_mutex_unlock(&job_mutex);
|
|
|
|
/* ask for someone to attend it */
|
|
sem_post(job_sem);
|
|
}
|
|
}
|
|
|
|
|
|
void job_wait(xs_val **job)
|
|
/* waits for an available job */
|
|
{
|
|
*job = NULL;
|
|
|
|
if (sem_wait(job_sem) == 0) {
|
|
/* lock the mutex */
|
|
pthread_mutex_lock(&job_mutex);
|
|
|
|
/* dequeue */
|
|
job_fifo_item *i = job_fifo_first;
|
|
|
|
if (i != NULL) {
|
|
job_fifo_first = i->next;
|
|
|
|
if (job_fifo_first == NULL)
|
|
job_fifo_last = NULL;
|
|
|
|
*job = i->job;
|
|
xs_free(i);
|
|
|
|
p_state->job_fifo_size--;
|
|
}
|
|
|
|
/* unlock the mutex */
|
|
pthread_mutex_unlock(&job_mutex);
|
|
}
|
|
}
|
|
|
|
|
|
static void *job_thread(void *arg)
|
|
/* job thread */
|
|
{
|
|
int pid = (int)(uintptr_t)arg;
|
|
|
|
srv_debug(1, xs_fmt("job thread %d started", pid));
|
|
|
|
for (;;) {
|
|
xs *job = NULL;
|
|
|
|
p_state->th_state[pid] = THST_WAIT;
|
|
|
|
job_wait(&job);
|
|
|
|
if (job == NULL) /* corrupted message? */
|
|
continue;
|
|
|
|
if (xs_type(job) == XSTYPE_FALSE) /* special message: exit */
|
|
break;
|
|
else
|
|
if (xs_type(job) == XSTYPE_DATA) {
|
|
/* it's a socket */
|
|
FILE *f = NULL;
|
|
|
|
p_state->th_state[pid] = THST_IN;
|
|
|
|
xs_data_get(&f, job);
|
|
|
|
if (f != NULL)
|
|
httpd_connection(f);
|
|
}
|
|
else {
|
|
/* it's a q_item */
|
|
p_state->th_state[pid] = THST_QUEUE;
|
|
|
|
process_queue_item(job);
|
|
}
|
|
}
|
|
|
|
p_state->th_state[pid] = THST_STOP;
|
|
|
|
srv_debug(1, xs_fmt("job thread %d stopped", pid));
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/* background thread sleep control */
|
|
static pthread_mutex_t sleep_mutex;
|
|
static pthread_cond_t sleep_cond;
|
|
|
|
static void *background_thread(void *arg)
|
|
/* background thread (queue management and other things) */
|
|
{
|
|
time_t purge_time;
|
|
|
|
(void)arg;
|
|
|
|
/* first purge time */
|
|
purge_time = time(NULL) + 10 * 60;
|
|
|
|
srv_log(xs_fmt("background thread started"));
|
|
|
|
while (p_state->srv_running) {
|
|
time_t t;
|
|
int cnt = 0;
|
|
|
|
p_state->th_state[0] = THST_QUEUE;
|
|
|
|
{
|
|
xs *list = user_list();
|
|
char *p, *uid;
|
|
|
|
/* process queues for all users */
|
|
p = list;
|
|
while (xs_list_iter(&p, &uid)) {
|
|
snac snac;
|
|
|
|
if (user_open(&snac, uid)) {
|
|
cnt += process_user_queue(&snac);
|
|
user_free(&snac);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* global queue */
|
|
cnt += process_queue();
|
|
|
|
/* time to purge? */
|
|
if ((t = time(NULL)) > purge_time) {
|
|
/* next purge time is tomorrow */
|
|
purge_time = t + 24 * 60 * 60;
|
|
|
|
xs *q_item = xs_dict_new();
|
|
q_item = xs_dict_append(q_item, "type", "purge");
|
|
job_post(q_item, 0);
|
|
}
|
|
|
|
if (cnt == 0) {
|
|
/* sleep 3 seconds */
|
|
|
|
p_state->th_state[0] = THST_WAIT;
|
|
|
|
#ifdef USE_POLL_FOR_SLEEP
|
|
poll(NULL, 0, 3 * 1000);
|
|
#else
|
|
struct timespec ts;
|
|
|
|
clock_gettime(CLOCK_REALTIME, &ts);
|
|
ts.tv_sec += 3;
|
|
|
|
pthread_mutex_lock(&sleep_mutex);
|
|
while (pthread_cond_timedwait(&sleep_cond, &sleep_mutex, &ts) == 0);
|
|
pthread_mutex_unlock(&sleep_mutex);
|
|
#endif
|
|
}
|
|
}
|
|
|
|
p_state->th_state[0] = THST_STOP;
|
|
|
|
srv_log(xs_fmt("background thread stopped"));
|
|
|
|
return NULL;
|
|
}
|
|
|
|
|
|
void term_handler(int s)
|
|
{
|
|
(void)s;
|
|
|
|
longjmp(on_break, 1);
|
|
}
|
|
|
|
|
|
srv_state *srv_state_op(xs_str **fname, int op)
|
|
/* opens or deletes the shared memory object */
|
|
{
|
|
int fd;
|
|
srv_state *ss = NULL;
|
|
|
|
if (*fname == NULL)
|
|
*fname = xs_fmt("/%s_snac_state", xs_dict_get(srv_config, "host"));
|
|
|
|
switch (op) {
|
|
case 0: /* open for writing */
|
|
if ((fd = shm_open(*fname, O_CREAT | O_RDWR, 0666)) != -1) {
|
|
ftruncate(fd, sizeof(*ss));
|
|
|
|
if ((ss = mmap(0, sizeof(*ss), PROT_READ | PROT_WRITE,
|
|
MAP_SHARED, fd, 0)) == MAP_FAILED)
|
|
ss = NULL;
|
|
|
|
close(fd);
|
|
}
|
|
|
|
if (ss == NULL) {
|
|
/* shared memory error: just create a plain structure */
|
|
srv_log(xs_fmt("warning: shm object error (%s)", strerror(errno)));
|
|
ss = malloc(sizeof(*ss));
|
|
}
|
|
|
|
/* init structure */
|
|
*ss = (srv_state){0};
|
|
ss->s_size = sizeof(*ss);
|
|
|
|
break;
|
|
|
|
case 1: /* open for reading */
|
|
if ((fd = shm_open(*fname, O_RDONLY, 0666)) != -1) {
|
|
if ((ss = mmap(0, sizeof(*ss), PROT_READ, MAP_SHARED, fd, 0)) == MAP_FAILED)
|
|
ss = NULL;
|
|
|
|
close(fd);
|
|
}
|
|
|
|
if (ss == NULL) {
|
|
/* shared memory error */
|
|
srv_log(xs_fmt("error: shm object error (%s) server not running?", strerror(errno)));
|
|
}
|
|
else
|
|
if (ss->s_size != sizeof(*ss)) {
|
|
srv_log(xs_fmt("error: struct size mismatch (%d != %d)",
|
|
ss->s_size, sizeof(*ss)));
|
|
|
|
munmap(ss, sizeof(*ss));
|
|
|
|
ss = NULL;
|
|
}
|
|
|
|
break;
|
|
|
|
case 2: /* unlink */
|
|
if (*fname)
|
|
shm_unlink(*fname);
|
|
|
|
break;
|
|
}
|
|
|
|
return ss;
|
|
}
|
|
|
|
|
|
void httpd(void)
|
|
/* starts the server */
|
|
{
|
|
const char *address;
|
|
const char *port;
|
|
int rs;
|
|
pthread_t threads[MAX_THREADS] = {0};
|
|
int n;
|
|
xs *sem_name = NULL;
|
|
xs *shm_name = NULL;
|
|
sem_t anon_job_sem;
|
|
|
|
address = xs_dict_get(srv_config, "address");
|
|
port = xs_number_str(xs_dict_get(srv_config, "port"));
|
|
|
|
if ((rs = xs_socket_server(address, port)) == -1) {
|
|
srv_log(xs_fmt("cannot bind socket to %s:%s", address, port));
|
|
return;
|
|
}
|
|
|
|
/* setup the server stat structure */
|
|
p_state = srv_state_op(&shm_name, 0);
|
|
|
|
p_state->srv_start_time = time(NULL);
|
|
|
|
p_state->use_fcgi = xs_type(xs_dict_get(srv_config, "fastcgi")) == XSTYPE_TRUE;
|
|
|
|
p_state->srv_running = 1;
|
|
|
|
signal(SIGPIPE, SIG_IGN);
|
|
signal(SIGTERM, term_handler);
|
|
signal(SIGINT, term_handler);
|
|
|
|
srv_log(xs_fmt("httpd%s start %s:%s %s", p_state->use_fcgi ? " (FastCGI)" : "",
|
|
address, port, USER_AGENT));
|
|
|
|
/* show the number of usable file descriptors */
|
|
struct rlimit r;
|
|
getrlimit(RLIMIT_NOFILE, &r);
|
|
srv_debug(0, xs_fmt("available (rlimit) fds: %d (cur) / %d (max)",
|
|
(int) r.rlim_cur, (int) r.rlim_max));
|
|
|
|
/* initialize the job control engine */
|
|
pthread_mutex_init(&job_mutex, NULL);
|
|
sem_name = xs_fmt("/job_%d", getpid());
|
|
job_sem = sem_open(sem_name, O_CREAT, 0644, 0);
|
|
|
|
if (job_sem == NULL) {
|
|
/* error opening a named semaphore; try with an anonymous one */
|
|
if (sem_init(&anon_job_sem, 0, 0) != -1)
|
|
job_sem = &anon_job_sem;
|
|
}
|
|
|
|
if (job_sem == NULL) {
|
|
srv_log(xs_fmt("fatal error: cannot create semaphore -- cannot continue"));
|
|
return;
|
|
}
|
|
|
|
/* initialize sleep control */
|
|
pthread_mutex_init(&sleep_mutex, NULL);
|
|
pthread_cond_init(&sleep_cond, NULL);
|
|
|
|
p_state->n_threads = xs_number_get(xs_dict_get(srv_config, "num_threads"));
|
|
|
|
#ifdef _SC_NPROCESSORS_ONLN
|
|
if (p_state->n_threads == 0) {
|
|
/* get number of CPUs on the machine */
|
|
p_state->n_threads = sysconf(_SC_NPROCESSORS_ONLN);
|
|
}
|
|
#endif
|
|
|
|
if (p_state->n_threads < 4)
|
|
p_state->n_threads = 4;
|
|
|
|
if (p_state->n_threads > MAX_THREADS)
|
|
p_state->n_threads = MAX_THREADS;
|
|
|
|
srv_debug(0, xs_fmt("using %d threads", p_state->n_threads));
|
|
|
|
/* thread #0 is the background thread */
|
|
pthread_create(&threads[0], NULL, background_thread, NULL);
|
|
|
|
/* the rest of threads are for job processing */
|
|
char *ptr = (char *) 0x1;
|
|
for (n = 1; n < p_state->n_threads; n++)
|
|
pthread_create(&threads[n], NULL, job_thread, ptr++);
|
|
|
|
if (setjmp(on_break) == 0) {
|
|
for (;;) {
|
|
FILE *f = xs_socket_accept(rs);
|
|
|
|
if (f != NULL) {
|
|
xs *job = xs_data_new(&f, sizeof(FILE *));
|
|
job_post(job, 1);
|
|
}
|
|
else
|
|
break;
|
|
}
|
|
}
|
|
|
|
p_state->srv_running = 0;
|
|
|
|
/* send as many exit jobs as working threads */
|
|
for (n = 1; n < p_state->n_threads; n++)
|
|
job_post(xs_stock(XSTYPE_FALSE), 0);
|
|
|
|
/* wait for all the threads to exit */
|
|
for (n = 0; n < p_state->n_threads; n++)
|
|
pthread_join(threads[n], NULL);
|
|
|
|
sem_close(job_sem);
|
|
sem_unlink(sem_name);
|
|
|
|
srv_state_op(&shm_name, 2);
|
|
|
|
xs *uptime = xs_str_time_diff(time(NULL) - p_state->srv_start_time);
|
|
|
|
srv_log(xs_fmt("httpd%s stop %s:%s (run time: %s)",
|
|
p_state->use_fcgi ? " (FastCGI)" : "",
|
|
address, port, uptime));
|
|
}
|