2024.11.27 - [up_ranoz] Modify download link to not use the upload url ticket link
# 2024.11.26 - [filehaus] Handle "404 Not found" on first instance # 2024.11.25 - [up_moocloud / moocloud] Add moocloud.ch as an upload and download host # 2024.11.24 - [uploadhive] Handle "Error creating download link" response -- do not mark Removed # 2024.11.23 - [filehaus] Use tor_curl_request_extended for head / get for filehaus urls # 2024.11.23 - [mad] Make tor_curl_request_extended a random timeout between 30-60 seconds # 2024.11.22 - [up_quax, quax] Add qu.ax as an upload and download host # 2024.11.21 - [filedot] Fix check for post filename # 2024.11.20 - [gofile] Handle parsing parent gofile url into multiple download urls # (still needs updating to handle child urls gofile.io/download/web/<guid>/file) # 2024.11.19 - [mad] Add updateUrlDownload function to handle updating a url # (ie. parent gofile url with children urls)
This commit is contained in:
parent
fa83163a58
commit
ec7d121c0e
12 changed files with 656 additions and 34 deletions
|
|
@ -97,21 +97,17 @@ fh_FetchFileInfo() {
|
|||
tor_identity="${RANDOM}"
|
||||
CLEANSTRING=${remote_url//[^a-zA-Z0-9]/}
|
||||
trap "rm -f ${WorkDir}/.flocks/${CLEANSTRING}; echo ""; tput cnorm; exit" 0 1 2 3 6 15
|
||||
file_header=$(tor_curl_request --insecure -L --head -s --referer "${remote_url//\.org/\.cc}" "$download_url")
|
||||
file_header=$(tor_curl_request_extended --insecure -L --head -s --referer "${remote_url//\.org/\.cc}" "$download_url")
|
||||
if [ "${DebugAllEnabled}" == "true" ] ; then
|
||||
debugHtml "${remote_url##*/}" "fh_head$j" "download_url: ${download_url}"$'\n'"${file_header}"
|
||||
fi
|
||||
if [ ! -z "$file_header" ] ; then
|
||||
if grep -Eqi '404 Not Found' <<< "${file_header}" ; then
|
||||
if [ $j == $maxfetchretries ] ; then
|
||||
printf "\\n"
|
||||
echo -e "${RED}| The file has been removed (404).${NC}"
|
||||
removedDownload "${remote_url}"
|
||||
exitDownloadNotAvailable=true
|
||||
return 1
|
||||
else
|
||||
continue
|
||||
fi
|
||||
fi
|
||||
if ! grep -Eqi '200|content-length' <<< "${file_header}" ; then
|
||||
if [ $j == $maxfetchretries ] ; then
|
||||
|
|
@ -194,9 +190,9 @@ fh_GetFile() {
|
|||
CLEANSTRING=${remote_url//[^a-zA-Z0-9]/}
|
||||
trap "rm -f ${WorkDir}/.flocks/${CLEANSTRING}; rm -f $flockDownload; echo ""; tput cnorm; exit" 0 1 2 3 6 15
|
||||
if [ "${RateMonitorEnabled}" == "true" ]; then
|
||||
tor_curl_request --insecure --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval --referer "$download_url" "$download_url" --continue-at - --output "$file_path"
|
||||
tor_curl_request_extended --insecure --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval --referer "$download_url" "$download_url" --continue-at - --output "$file_path"
|
||||
else
|
||||
tor_curl_request --insecure --referer "$download_url" "$download_url" --continue-at - --output "$file_path"
|
||||
tor_curl_request_extended --insecure --referer "$download_url" "$download_url" --continue-at - --output "$file_path"
|
||||
fi
|
||||
received_file_size=0
|
||||
if [ -f "$file_path" ] ; then
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue