faviconslib/favicon.php

228 lines
8.0 KiB
PHP
Raw Permalink Normal View History

2014-07-18 00:28:34 +02:00
<?php
/** Favicons Lib
* ------------
* @copyright SODAWARE License (See below)
* @brief Simple lib to try to get favicons from URLs.
*/
/* LICENSE
* --------------------------------------------------------------------------------
* "THE NO-ALCOHOL BEER-WARE LICENSE" (Revision 42):
* Phyks (webmaster@phyks.me) wrote this file. As long as you retain this notice you
* can do whatever you want with this stuff (and you can also do whatever you want
* with this stuff without retaining it, but that's not cool...). If we meet some
* day, and you think this stuff is worth it, you can buy me a <del>beer</del> soda
* in return.
* Phyks
* ---------------------------------------------------------------------------------
*/
/**
* Try to get the favicon associated with some URLs, by parsing the header and
* trying to get the file favicon.ico at the root of the server
*
* @param an array $urls of URLs
* @return an array {'favicons', 'errors'}. `errors` is an array of URLs for which there could not be any fetched favicon. `favicons` is an array with URLs as keys and an array of favicon urls and sizes ({favicon_url, size}, associative array).
*/
function getFavicon($urls) {
$favicons = array();
$errors = array();
// Convert array to the good format for curl downloader
$curl_urls = array();
foreach($urls as $url) {
if (endswith($url, '.html')) { // Only check html files using first method
$curl_urls[] = array('url'=>$url);
}
else {
$errors[] = $url;
}
2014-07-18 00:28:34 +02:00
}
$contents = curl_downloader($curl_urls);
foreach($contents['status_codes'] as $url=>$status) {
if($status != 200) {
$errors[] = $url;
}
}
foreach($contents['results'] as $url=>$content) {
2014-07-30 23:36:58 +02:00
$content = substr($content, 0, strpos($content, '</head>')).'</head></html>'; // We don't need the full page, just the <head>
2014-07-18 00:28:34 +02:00
$html = new DOMDocument();
$html->strictErrorChecking = false;
2014-07-18 23:13:33 +02:00
$success = @$html->loadHTML($content);
if($success === false) {
continue;
}
2014-07-18 00:28:34 +02:00
$xml = simplexml_import_dom($html);
2014-07-18 23:13:33 +02:00
if($xml === false) {
continue;
}
2014-07-18 00:28:34 +02:00
// Try to fetch the favicon URL from the <head> tag
foreach($xml->head->children() as $head_tag) {
if($head_tag->getName() != 'link') {
continue;
}
$go_next_tag = false;
foreach($head_tag->attributes() as $key=>$attribute) {
if($go_next_tag || $key != 'rel') {
continue;
}
if(strstr((string) $attribute, 'icon')) {
2014-08-29 12:46:40 +02:00
$tmp = $head_tag->attributes();
if(isset($tmp['sizes'])) {
2014-08-29 16:14:09 +02:00
$sizes = (string) $tmp['sizes'];
2014-07-18 00:28:34 +02:00
}
else {
$sizes = '';
}
2014-07-18 23:22:48 +02:00
$favicons[$url][] = array(
2014-08-29 16:14:09 +02:00
'favicon_url'=>(string) $tmp['href'],
2014-07-18 00:28:34 +02:00
'sizes'=>$sizes
);
$go_next_tag = true;
}
}
}
}
// Add to errors the URLs without any favicons associated
$favicons_keys = array_keys($favicons);
foreach($contents['results'] as $url=>$content) {
if(!in_array($url, $favicons_keys)) {
$errors[] = $url;
}
}
// Check for errorred feeds wether the favicon.ico file at the root exists
$second_try = array();
foreach ($errors as $url) {
2014-07-18 18:53:43 +02:00
$parsed_url = parse_url(trim($url));
$second_try_url = "";
if(isset($parsed_url['scheme'])) {
$second_try_url .= $parsed_url['scheme'].'://';
2014-07-18 18:53:43 +02:00
}
if(isset($parsed_url['host'])) {
$second_try_url .= $parsed_url['host'];
}
if(isset($parsed_url['port'])) {
$second_try_url .= $parsed_url['port'];
}
if(isset($parsed_url['user'])) {
$second_try_url .= $parsed_url['user'];
}
if(isset($parsed_url['pass'])) {
$second_try_url .= $parsed_url['pass'];
}
2014-07-18 00:28:34 +02:00
$second_try[] = array(
'input_url'=>$url,
2014-07-18 18:53:43 +02:00
'url'=>$second_try_url . '/favicon.ico'
2014-07-18 00:28:34 +02:00
);
}
$second_try_curl = curl_downloader($second_try, false);
$errors = array();
foreach($second_try as $tested_url) {
$status_code = (int) $second_try_curl['status_codes'][$tested_url['url']];
if ($status_code >= 200 && $status_code < 400) {
2014-07-18 23:22:48 +02:00
$favicons[$tested_url['input_url']][] = array(
2014-07-18 00:28:34 +02:00
'favicon_url'=>$tested_url['url'],
'sizes'=>''
);
}
else {
$errors[] = $tested_url['input_url'];
}
}
return array('favicons'=>$favicons, 'errors'=>$errors);
}
/**
* Downloads all the urls in the array $urls and returns an array with the results and the http status_codes.
*
* Mostly inspired by blogotext by timovn : https://github.com/timovn/blogotext/blob/master/inc/fich.php
*
* @todo If open_basedir or safe_mode, Curl will not follow redirections :
* https://stackoverflow.com/questions/24687145/curlopt-followlocation-and-curl-multi-and-safe-mode
*
* @param an array $urls of associative arrays {'url', 'post'} for each URL. 'post' is a JSON array of data to send _via_ POST.
* @return an array {'results', 'status_code'}, results being an array of the retrieved contents, indexed by URLs, and 'status_codes' being an array of status_code, indexed by URL.
*/
function curl_downloader($urls, $fetch_content=true) {
$chunks = array_chunk($urls, 40, true); // Chunks of 40 urls because curl has problems with too big "multi" requests
$results = array();
$status_codes = array();
if (ini_get('open_basedir') == '' && ini_get('safe_mode') === false) { // Disable followlocation option if this is activated, to avoid warnings
$follow_redirect = true;
}
else {
$follow_redirect = false;
}
foreach ($chunks as $chunk) {
$multihandler = curl_multi_init();
$handlers = array();
$total_feed_chunk = count($chunk) + count($results);
foreach ($chunk as $i=>$url_array) {
$url = $url_array['url'];
set_time_limit(20); // Reset max execution time
$handlers[$i] = curl_init($url);
curl_setopt_array($handlers[$i], array(
CURLOPT_RETURNTRANSFER => TRUE,
CURLOPT_CONNECTTIMEOUT => 10,
CURLOPT_TIMEOUT => 15,
CURLOPT_FOLLOWLOCATION => $follow_redirect,
CURLOPT_MAXREDIRS => 5,
CURLOPT_USERAGENT => $_SERVER['HTTP_USER_AGENT'], // Add a user agent to prevent problems with some feeds
CURLOPT_HEADER => $fetch_content ? FALSE : TRUE,
CURLOPT_NOBODY => $fetch_content ? FALSE : TRUE,
));
if (!empty($url_array['post'])) {
curl_setopt($handlers[$i], CURLOPT_POST, true);
curl_setopt($handlers[$i], CURLOPT_POSTFIELDS, json_decode($url_array['post'], true));
}
curl_multi_add_handle($multihandler, $handlers[$i]);
}
do {
curl_multi_exec($multihandler, $active);
curl_multi_select($multihandler);
} while ($active > 0);
foreach ($chunk as $i=>$url_array) {
$url = $url_array['url'];
$results[$url] = curl_multi_getcontent($handlers[$i]);
$status_codes[$url] = curl_getinfo($handlers[$i], CURLINFO_HTTP_CODE);
curl_multi_remove_handle($multihandler, $handlers[$i]);
curl_close($handlers[$i]);
}
curl_multi_close($multihandler);
}
return array('results'=>$results, 'status_codes'=>$status_codes);
}
/**
* Check that $haystack ends with $needle.
*/
function endswith($haystack, $needle) {
$length = strlen($needle);
if ($length == 0) {
return true;
}
return (substr($haystack, -$length) === $needle);
}