-const char cgisimple_rcs[] = "$Id: cgisimple.c,v 1.75 2008/04/27 13:52:52 fabiankeil Exp $";
+const char cgisimple_rcs[] = "$Id: cgisimple.c,v 1.82 2008/05/10 20:01:47 fabiankeil Exp $";
/*********************************************************************
*
* File : $Source: /cvsroot/ijbswa/current/cgisimple.c,v $
*
* Revisions :
* $Log: cgisimple.c,v $
+ * Revision 1.82 2008/05/10 20:01:47 fabiankeil
+ * Fix an assertion that could erroneously
+ * trigger in case of memory shortage.
+ *
+ * Revision 1.81 2008/05/05 09:54:39 fabiankeil
+ * In cgi_show_url_info(), make sure ftp URLs are
+ * declared invalid. Also simplify the code that adds
+ * "http://" if no protocol has been specified.
+ *
+ * Revision 1.80 2008/05/04 16:18:32 fabiankeil
+ * Provide parse_http_url() with a third parameter to specify
+ * whether or not URLs without protocol are acceptable.
+ *
+ * Revision 1.79 2008/05/04 13:30:56 fabiankeil
+ * Streamline parse_http_url()'s prototype.
+ *
+ * Revision 1.78 2008/05/03 16:50:11 fabiankeil
+ * Leverage content_filters_enabled() in cgi_show_url_info().
+ *
+ * Revision 1.77 2008/05/02 09:47:48 fabiankeil
+ * In cgi_show_url_info, pass an initialized http structure
+ * to parse_http_url() as that will be required soonish and
+ * assert that https URLs are recognized correctly.
+ *
+ * Revision 1.76 2008/04/28 09:13:30 fabiankeil
+ * In load_file(), remember the error reason and fclose()
+ * and return later on instead of right away.
+ *
* Revision 1.75 2008/04/27 13:52:52 fabiankeil
* Move CGI file loading code into load_file() and
* add checks for unexpected errors.
url_param[0] = '\0';
}
}
- else if (url_param[0] != '\0')
+ else if ((url_param[0] != '\0') && (NULL == strstr(url_param, "://")))
{
- /*
- * Unknown prefix - assume http://
- */
- const size_t url_param_prefixed_size = 7 + 1 + strlen(url_param);
- char * url_param_prefixed = malloc(url_param_prefixed_size);
- if (NULL == url_param_prefixed)
+ /* No prefix - assume http:// */
+ char *url_param_prefixed = strdup("http://");
+
+ if (JB_ERR_OK != string_join(&url_param_prefixed, url_param))
{
- free(url_param);
free_map(exports);
return JB_ERR_MEMORY;
}
- strlcpy(url_param_prefixed, "http://", url_param_prefixed_size);
- strlcat(url_param_prefixed, url_param, url_param_prefixed_size);
- free(url_param);
url_param = url_param_prefixed;
}
return JB_ERR_MEMORY;
}
- err = parse_http_url(url_param, url_to_query, csp);
+ memset(url_to_query, '\0', sizeof(url_to_query));
+ err = parse_http_url(url_param, url_to_query, REQUIRE_PROTOCOL);
+ assert((err != JB_ERR_OK) || (url_to_query->ssl == !strncmp(url_param, "https://", 8)));
free(url_param);
* If zlib support is available, if no content filters
* are enabled or if the prevent-compression action is enabled,
* suppress the "compression could prevent filtering" warning.
- *
- * XXX: Change content_filters_enabled()'s prototype so we can
- * use it here.
*/
#ifndef FEATURE_ZLIB
- if ((list_is_empty(action->multi[ACTION_MULTI_FILTER])
- && !(action->flags & ACTION_DEANIMATE))
- || (action->flags & ACTION_NO_COMPRESSION))
+ if (!content_filters_enabled(action) ||
+ (action->flags & ACTION_NO_COMPRESSION))
#endif
{
if (!err) err = map_block_killer(exports, "filters-might-be-ineffective");