diff --git a/.audit/mad-audit-curl.log b/.audit/mad-audit-curl.log index 6399644..9220380 100755 --- a/.audit/mad-audit-curl.log +++ b/.audit/mad-audit-curl.log @@ -1,4 +1,4 @@ -DateTime: 24.12.26 +DateTime: 25.01.04 Files: ./hosts/1fichier.sh @@ -163,12 +163,12 @@ _________________________________________________________________________ ./hosts/anonfile.sh:186: response=$(tor_curl_request --insecure -L -s -X POST \ ./hosts/anonfile.sh:240: tor_curl_request --insecure -s "$captcha_img_url" --output "$tmp_captcha_img" ./hosts/anonfile.sh:340: response=$(tor_curl_request --insecure -L -s -X POST \ -./hosts/anonfile.sh:446: file_header=$(tor_curl_request -i -s --head \ -./hosts/anonfile.sh:549: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./hosts/anonfile.sh:551: tor_curl_request --insecure \ -./hosts/anonfile.sh:557: tor_curl_request --insecure \ -./hosts/anonfile.sh:564: tor_curl_request --insecure \ -./hosts/anonfile.sh:581: tor_curl_request --insecure \ +./hosts/anonfile.sh:451: file_header=$(tor_curl_request -i -s --head \ +./hosts/anonfile.sh:557: if [ "${UseTorCurlImpersonate}" == "true" ]; then +./hosts/anonfile.sh:559: tor_curl_request --insecure \ +./hosts/anonfile.sh:565: tor_curl_request --insecure \ +./hosts/anonfile.sh:572: tor_curl_request --insecure \ +./hosts/anonfile.sh:589: tor_curl_request --insecure \ ./hosts/anonsharing.sh:91: response=$(tor_curl_request --insecure -i -s \ ./hosts/anonsharing.sh:150: file_header=$(tor_curl_request --insecure --head -L -i -s \ ./hosts/anonsharing.sh:158: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -s -i \ @@ -219,14 +219,14 @@ _________________________________________________________________________ ./hosts/dailyuploads.sh:511: tor_curl_request --insecure \ ./hosts/dailyuploads.sh:528: tor_curl_request --insecure \ ./hosts/dashfile.sh:90: response=$(tor_curl_request --insecure -L -s -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ -./hosts/dashfile.sh:169: response=$(tor_curl_request --insecure -L -s -X POST \ -./hosts/dashfile.sh:297: response=$(tor_curl_request --insecure -L -s -X POST \ -./hosts/dashfile.sh:386: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url") -./hosts/dashfile.sh:484: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./hosts/dashfile.sh:486: tor_curl_request --insecure \ -./hosts/dashfile.sh:491: tor_curl_request --insecure \ +./hosts/dashfile.sh:177: response=$(tor_curl_request --insecure -L -s -X POST \ +./hosts/dashfile.sh:308: response=$(tor_curl_request --insecure -L -s -X POST \ +./hosts/dashfile.sh:397: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url") +./hosts/dashfile.sh:495: if [ "${UseTorCurlImpersonate}" == "true" ]; then ./hosts/dashfile.sh:497: tor_curl_request --insecure \ -./hosts/dashfile.sh:513: tor_curl_request --insecure \ +./hosts/dashfile.sh:502: tor_curl_request --insecure \ +./hosts/dashfile.sh:508: tor_curl_request --insecure \ +./hosts/dashfile.sh:524: tor_curl_request --insecure \ ./hosts/dataupload.sh:90: response=$(tor_curl_request --insecure -L -s -b "${dup_cookie_jar}" -c "${dup_cookie_jar}" "$remote_url") ./hosts/dataupload.sh:166: response=$(tor_curl_request --insecure -svo. -X POST \ ./hosts/dataupload.sh:234: file_header=$(tor_curl_request --insecure -L --head -s "$download_url") @@ -255,14 +255,14 @@ _________________________________________________________________________ ./hosts/examples/ExampleNewHost.sh:201: tor_curl_request --insecure --referer "$download_url" "$download_url" --continue-at - --output "$file_path" ./hosts/examples/up_example.sh:112: response=$(tor_curl_upload --insecure \ ./hosts/fileblade.sh:90: response=$(tor_curl_request --insecure -L -s -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" "$remote_url") -./hosts/fileblade.sh:167: response=$(tor_curl_request --insecure -L -s -X POST \ -./hosts/fileblade.sh:254: response=$(tor_curl_request --insecure -L -s -X POST \ -./hosts/fileblade.sh:323: file_header=$(tor_curl_request --insecure -L --head -s "$download_url") -./hosts/fileblade.sh:438: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./hosts/fileblade.sh:440: tor_curl_request --insecure -L \ -./hosts/fileblade.sh:444: tor_curl_request --insecure -L \ -./hosts/fileblade.sh:449: tor_curl_request --insecure \ -./hosts/fileblade.sh:464: tor_curl_request --insecure \ +./hosts/fileblade.sh:164: response=$(tor_curl_request --insecure -L -s -X POST \ +./hosts/fileblade.sh:266: response=$(tor_curl_request --insecure -L -s -X POST \ +./hosts/fileblade.sh:335: file_header=$(tor_curl_request --insecure -L --head -s "$download_url") +./hosts/fileblade.sh:450: if [ "${UseTorCurlImpersonate}" == "true" ]; then +./hosts/fileblade.sh:452: tor_curl_request --insecure -L \ +./hosts/fileblade.sh:456: tor_curl_request --insecure -L \ +./hosts/fileblade.sh:461: tor_curl_request --insecure \ +./hosts/fileblade.sh:476: tor_curl_request --insecure \ ./hosts/fileditch.sh:85: file_header=$(tor_curl_request --insecure --head -L -s "$download_url") ./hosts/fileditch.sh:176: tor_curl_request --insecure --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path" ./hosts/fileditch.sh:178: tor_curl_request --insecure "$download_url" --continue-at - --output "$file_path" @@ -282,9 +282,9 @@ _________________________________________________________________________ ./hosts/firestorage.sh:337: tor_curl_request --insecure -L -G "$download_url" --continue-at - --output "$file_path" ./hosts/gofile.sh:97: response=$(tor_curl_request --insecure -s -X POST \ ./hosts/gofile.sh:170: response=$(tor_curl_request --insecure -G -L -s \ -./hosts/gofile.sh:250: file_header=$(tor_curl_request --insecure -L --head -s \ -./hosts/gofile.sh:369: tor_curl_request --insecure -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ -./hosts/gofile.sh:383: tor_curl_request --insecure -G \ +./hosts/gofile.sh:258: file_header=$(tor_curl_request --insecure -L --head -s \ +./hosts/gofile.sh:377: tor_curl_request --insecure -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ +./hosts/gofile.sh:391: tor_curl_request --insecure -G \ ./hosts/hexload.sh:108: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download") ./hosts/hexload.sh:116: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download") ./hosts/hexload.sh:122: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download") @@ -452,119 +452,119 @@ _________________________________________________________________________ ./hosts/youdbox.sh:183: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url") ./hosts/youdbox.sh:276: tor_curl_request --insecure -L -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path" ./hosts/youdbox.sh:278: tor_curl_request --insecure -L -G "$download_url" --continue-at - --output "$file_path" -./mad.sh:100:UseTorCurlImpersonate=false -./mad.sh:391:tor_curl_request() { -./mad.sh:392: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:393: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" -./mad.sh:395: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" -./mad.sh:398:tor_curl_request_extended() { -./mad.sh:400: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:401: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@" -./mad.sh:403: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@" -./mad.sh:406:tor_curl_upload() { +./mad.sh:107:UseTorCurlImpersonate=false +./mad.sh:398:tor_curl_request() { +./mad.sh:399: if [ "${UseTorCurlImpersonate}" == "true" ]; then +./mad.sh:400: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" +./mad.sh:402: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" +./mad.sh:405:tor_curl_request_extended() { ./mad.sh:407: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:409: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval --compressed --globoff "$@" -./mad.sh:411: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" -./mad.sh:415: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@" -./mad.sh:417: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@" -./mad.sh:1361:install_curl_impersonate() { -./mad.sh:1363: echo -e "${BLUE}lwthiker curl_impersonate${NC} is the original dev, but it is relatively inactive." -./mad.sh:1364: echo -e "- Currently uses curl v8.1.1." -./mad.sh:1368: echo -e "${BLUE}lexiforest curl_impersonate${NC} is an active fork of curl_impersonate." -./mad.sh:1369: echo -e "+ Currently uses curl v8.7.1" -./mad.sh:1373: PS3='Please select which curl_impersonate to install: ' -./mad.sh:1381: install_curl_impersonate_lwthiker_orig -./mad.sh:1385: install_curl_impersonate_lexiforest_fork -./mad.sh:1395:install_curl_impersonate_lwthiker_orig() { -./mad.sh:1399: echo -e "${BLUE}lwthiker curl_impersonate${NC} is the original curl_impersonate." -./mad.sh:1400: echo -e "+ Currently uses curl v8.1.1, and has low activity for updates" -./mad.sh:1403: echo -e "${GREEN}| Fetching:${NC} latest ${BLUE}lwthiker curl_impersonate${NC} info from github...${NC}" -./mad.sh:1406: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest) -./mad.sh:1408: debugHtml "github" "lbf_inst_curlimp$j" "$response" -./mad.sh:1411: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response") -./mad.sh:1421: yes_or_no "Do you wish to download and extract curl_impersonate $latestTag (using tor+curl)?" && { -./mad.sh:1423: download_url='https://github.com/lwthiker/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz' -./mad.sh:1426: file_header=$(tor_curl_request --insecure --head -Ls "$download_url") -./mad.sh:1428: debugHtml "github" "head_inst_curlimp$j" "${file_header}" -./mad.sh:1476: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path" -./mad.sh:1505: echo -e "| Extracting curl_impersonate..." -./mad.sh:1507: rm -f "${ScriptDir}"/curl* -./mad.sh:1508: mv "$extract_location/curl-impersonate-ff" "${ScriptDir}/" -./mad.sh:1509: mv "$extract_location/curl_ff109" "${ScriptDir}/" -./mad.sh:1510: echo -e "${GREEN}| Done.${NC} Update ${BLUE}\"UseTorCurlImpersonate=true\"${NC} in script to use..." -./mad.sh:1518:install_curl_impersonate_lexiforest_fork() { -./mad.sh:1522: echo -e "${BLUE}lexiforest curl_impersonate${NC} is an active fork of curl_impersonate." -./mad.sh:1523: echo -e "+ Currently uses curl v8.7.1, and is patched for latest CVEs" -./mad.sh:1526: echo -e "${GREEN}| Fetching:${NC} latest ${BLUE}lexiforest curl_impersonate fork${NC} info from github...${NC}" -./mad.sh:1529: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest) -./mad.sh:1531: debugHtml "github" "lbf_inst_curlimp$j" "$response" -./mad.sh:1534: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response") -./mad.sh:1544: yes_or_no "Do you wish to download and extract curl_impersonate $latestTag (using tor+curl)?" && { -./mad.sh:1546: download_url='https://github.com/lexiforest/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz' -./mad.sh:1549: file_header=$(tor_curl_request --insecure --head -Ls "$download_url") -./mad.sh:1551: debugHtml "github" "head_inst_curlimp$j" "${file_header}" -./mad.sh:1599: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path" -./mad.sh:1628: echo -e "| Extracting curl_impersonate..." -./mad.sh:1630: rm -f "${ScriptDir}"/curl* -./mad.sh:1631: mv "$extract_location/curl-impersonate-chrome" "${ScriptDir}/" -./mad.sh:1632: mv "$extract_location/curl_chrome131" "${ScriptDir}/" -./mad.sh:1633: echo -e "${GREEN}| Done.${NC} Update ${BLUE}\"UseTorCurlImpersonate=true\"${NC} in script to use..." -./mad.sh:1795: echo -e ":${NC} ${GREEN}MAD${PINK} Audit${NC} : Reports usage of http & curl in scripts${PINK}${BLD} :" -./mad.sh:1803: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl') -./mad.sh:1804: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') -./mad.sh:1813: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})" -./mad.sh:1815: echo -e "$maud_curl" -./mad.sh:1817: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" -./mad.sh:1819: echo -e "$maud_torcurl" -./mad.sh:1831: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl') -./mad.sh:1832: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') -./mad.sh:1841: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl \"${NC})" -./mad.sh:1843: echo -e "$maud_curl" -./mad.sh:1845: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" -./mad.sh:1847: echo -e "$maud_torcurl" -./mad.sh:1853: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl') -./mad.sh:1854: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') -./mad.sh:1863: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})" -./mad.sh:1865: echo -e "$maud_curl" -./mad.sh:1867: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" -./mad.sh:1869: echo -e "$maud_torcurl" -./mad.sh:2816: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:2817: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n" -./mad.sh:2819: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n" -./mad.sh:2991: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:2992: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n" -./mad.sh:2994: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n" -./mad.sh:3192: file_header=$(tor_curl_request --insecure -m 8 -s -D - -o /dev/null \ -./mad.sh:3199: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -L -s -i "$download_url" | -./mad.sh:3327: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --output "$file_path" -./mad.sh:3371: tor_curl_request --insecure -L --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path" -./mad.sh:3373: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --continue-at - --output "$file_path" -./mad.sh:3571: response=$(tor_curl_upload --insecure -i \ +./mad.sh:408: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@" +./mad.sh:410: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@" +./mad.sh:413:tor_curl_upload() { +./mad.sh:414: if [ "${UseTorCurlImpersonate}" == "true" ]; then +./mad.sh:416: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval --compressed --globoff "$@" +./mad.sh:418: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" +./mad.sh:422: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@" +./mad.sh:424: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@" +./mad.sh:1368:install_curl_impersonate() { +./mad.sh:1370: echo -e "${BLUE}lwthiker curl_impersonate${NC} is the original dev, but it is relatively inactive." +./mad.sh:1371: echo -e "- Currently uses curl v8.1.1." +./mad.sh:1375: echo -e "${BLUE}lexiforest curl_impersonate${NC} is an active fork of curl_impersonate." +./mad.sh:1376: echo -e "+ Currently uses curl v8.7.1" +./mad.sh:1380: PS3='Please select which curl_impersonate to install: ' +./mad.sh:1388: install_curl_impersonate_lwthiker_orig +./mad.sh:1392: install_curl_impersonate_lexiforest_fork +./mad.sh:1402:install_curl_impersonate_lwthiker_orig() { +./mad.sh:1406: echo -e "${BLUE}lwthiker curl_impersonate${NC} is the original curl_impersonate." +./mad.sh:1407: echo -e "+ Currently uses curl v8.1.1, and has low activity for updates" +./mad.sh:1410: echo -e "${GREEN}| Fetching:${NC} latest ${BLUE}lwthiker curl_impersonate${NC} info from github...${NC}" +./mad.sh:1413: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest) +./mad.sh:1415: debugHtml "github" "lbf_inst_curlimp$j" "$response" +./mad.sh:1418: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response") +./mad.sh:1428: yes_or_no "Do you wish to download and extract curl_impersonate $latestTag (using tor+curl)?" && { +./mad.sh:1430: download_url='https://github.com/lwthiker/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz' +./mad.sh:1433: file_header=$(tor_curl_request --insecure --head -Ls "$download_url") +./mad.sh:1435: debugHtml "github" "head_inst_curlimp$j" "${file_header}" +./mad.sh:1483: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path" +./mad.sh:1512: echo -e "| Extracting curl_impersonate..." +./mad.sh:1514: rm -f "${ScriptDir}"/curl* +./mad.sh:1515: mv "$extract_location/curl-impersonate-ff" "${ScriptDir}/" +./mad.sh:1516: mv "$extract_location/curl_ff109" "${ScriptDir}/" +./mad.sh:1517: echo -e "${GREEN}| Done.${NC} Update ${BLUE}\"UseTorCurlImpersonate=true\"${NC} in script to use..." +./mad.sh:1525:install_curl_impersonate_lexiforest_fork() { +./mad.sh:1529: echo -e "${BLUE}lexiforest curl_impersonate${NC} is an active fork of curl_impersonate." +./mad.sh:1530: echo -e "+ Currently uses curl v8.7.1, and is patched for latest CVEs" +./mad.sh:1533: echo -e "${GREEN}| Fetching:${NC} latest ${BLUE}lexiforest curl_impersonate fork${NC} info from github...${NC}" +./mad.sh:1536: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest) +./mad.sh:1538: debugHtml "github" "lbf_inst_curlimp$j" "$response" +./mad.sh:1541: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response") +./mad.sh:1551: yes_or_no "Do you wish to download and extract curl_impersonate $latestTag (using tor+curl)?" && { +./mad.sh:1553: download_url='https://github.com/lexiforest/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz' +./mad.sh:1556: file_header=$(tor_curl_request --insecure --head -Ls "$download_url") +./mad.sh:1558: debugHtml "github" "head_inst_curlimp$j" "${file_header}" +./mad.sh:1606: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path" +./mad.sh:1635: echo -e "| Extracting curl_impersonate..." +./mad.sh:1637: rm -f "${ScriptDir}"/curl* +./mad.sh:1638: mv "$extract_location/curl-impersonate-chrome" "${ScriptDir}/" +./mad.sh:1639: mv "$extract_location/curl_chrome131" "${ScriptDir}/" +./mad.sh:1640: echo -e "${GREEN}| Done.${NC} Update ${BLUE}\"UseTorCurlImpersonate=true\"${NC} in script to use..." +./mad.sh:1802: echo -e ":${NC} ${GREEN}MAD${PINK} Audit${NC} : Reports usage of http & curl in scripts${PINK}${BLD} :" +./mad.sh:1810: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl') +./mad.sh:1811: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') +./mad.sh:1820: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})" +./mad.sh:1822: echo -e "$maud_curl" +./mad.sh:1824: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" +./mad.sh:1826: echo -e "$maud_torcurl" +./mad.sh:1838: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl') +./mad.sh:1839: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') +./mad.sh:1848: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl \"${NC})" +./mad.sh:1850: echo -e "$maud_curl" +./mad.sh:1852: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" +./mad.sh:1854: echo -e "$maud_torcurl" +./mad.sh:1860: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl') +./mad.sh:1861: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') +./mad.sh:1870: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})" +./mad.sh:1872: echo -e "$maud_curl" +./mad.sh:1874: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" +./mad.sh:1876: echo -e "$maud_torcurl" +./mad.sh:2823: if [ "${UseTorCurlImpersonate}" == "true" ]; then +./mad.sh:2824: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n" +./mad.sh:2826: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n" +./mad.sh:2998: if [ "${UseTorCurlImpersonate}" == "true" ]; then +./mad.sh:2999: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n" +./mad.sh:3001: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n" +./mad.sh:3199: file_header=$(tor_curl_request --insecure -m 8 -s -D - -o /dev/null \ +./mad.sh:3206: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -L -s -i "$download_url" | +./mad.sh:3334: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --output "$file_path" +./mad.sh:3378: tor_curl_request --insecure -L --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path" +./mad.sh:3380: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --continue-at - --output "$file_path" ./mad.sh:3578: response=$(tor_curl_upload --insecure -i \ -./mad.sh:3649:if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:3650: curl_impersonate=() -./mad.sh:3651: readarray -d $'' arrFiles < <(find "$ScriptDir" -maxdepth 1 -name "curl_*" -printf '%p\n' | sort -Vk1) -./mad.sh:3652: bFoundCurlHeader=false -./mad.sh:3656: curl_impersonate=($fil) -./mad.sh:3657: bFoundCurlHeader=true -./mad.sh:3661: if [ "$bFoundCurlHeader" == "false" ]; then -./mad.sh:3662: echo -e "${RED}[ERROR] Missing dependency \"curl-impersonate\"!${NC}" -./mad.sh:3665: echo -e "You'll need to download ${GREEN}\"curl-impersonate\"${NC}." -./mad.sh:3668: echo -e "The latest binary can be obtained on GitHub, search for \"curl-impersonate\"" -./mad.sh:3670: echo -e " 1. Visit the page of curl-impersonate and add \"/releases/latest/\" at end of URL." -./mad.sh:3674: echo -e " 4. Download archive ${GREEN}\"curl-impersonate-vX.Y.Z.x86_64-linux-gnu.tar.gz\"${YELLOW}." -./mad.sh:3675: echo -e " 5. Extract files ${GREEN}\"curl-impersonate-ff\"${NC} and ${GREEN}\"curl_ff109\"${NC} next to this script." -./mad.sh:3678: echo -e "run $0 install_curl_impersonate\\n" -./mad.sh:3680: yes_or_no "Do you wish to download and extract latest curl_impersonate (using tor+curl)?" && { -./mad.sh:3681: UseTorCurlImpersonate=false -./mad.sh:3682: install_curl_impersonate -./mad.sh:3766: echo -e "[${YELLOW}Install curl_impersonate${NC}]: Downloads the latest binary for curl_impersonate from github repo (3 choices)" -./mad.sh:3767: printf " %s install_curl_impersonate\\n" "$0" -./mad.sh:3845:elif [[ "$arg1" == "install_curl_impersonate" ]]; then -./mad.sh:3846: install_curl_impersonate -./mad.sh:3877:if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:3878: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n" -./mad.sh:3880: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n" +./mad.sh:3585: response=$(tor_curl_upload --insecure -i \ +./mad.sh:3656:if [ "${UseTorCurlImpersonate}" == "true" ]; then +./mad.sh:3657: curl_impersonate=() +./mad.sh:3658: readarray -d $'' arrFiles < <(find "$ScriptDir" -maxdepth 1 -name "curl_*" -printf '%p\n' | sort -Vk1) +./mad.sh:3659: bFoundCurlHeader=false +./mad.sh:3663: curl_impersonate=($fil) +./mad.sh:3664: bFoundCurlHeader=true +./mad.sh:3668: if [ "$bFoundCurlHeader" == "false" ]; then +./mad.sh:3669: echo -e "${RED}[ERROR] Missing dependency \"curl-impersonate\"!${NC}" +./mad.sh:3672: echo -e "You'll need to download ${GREEN}\"curl-impersonate\"${NC}." +./mad.sh:3675: echo -e "The latest binary can be obtained on GitHub, search for \"curl-impersonate\"" +./mad.sh:3677: echo -e " 1. Visit the page of curl-impersonate and add \"/releases/latest/\" at end of URL." +./mad.sh:3681: echo -e " 4. Download archive ${GREEN}\"curl-impersonate-vX.Y.Z.x86_64-linux-gnu.tar.gz\"${YELLOW}." +./mad.sh:3682: echo -e " 5. Extract files ${GREEN}\"curl-impersonate-ff\"${NC} and ${GREEN}\"curl_ff109\"${NC} next to this script." +./mad.sh:3685: echo -e "run $0 install_curl_impersonate\\n" +./mad.sh:3687: yes_or_no "Do you wish to download and extract latest curl_impersonate (using tor+curl)?" && { +./mad.sh:3688: UseTorCurlImpersonate=false +./mad.sh:3689: install_curl_impersonate +./mad.sh:3773: echo -e "[${YELLOW}Install curl_impersonate${NC}]: Downloads the latest binary for curl_impersonate from github repo (3 choices)" +./mad.sh:3774: printf " %s install_curl_impersonate\\n" "$0" +./mad.sh:3852:elif [[ "$arg1" == "install_curl_impersonate" ]]; then +./mad.sh:3853: install_curl_impersonate +./mad.sh:3884:if [ "${UseTorCurlImpersonate}" == "true" ]; then +./mad.sh:3885: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n" +./mad.sh:3887: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n" ./plugins/pjscloud.sh:44: if [ "${UseTorCurlImpersonate}" == "true" ]; then ./plugins/pjscloud.sh:45: response=$("${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" \ ./plugins/pjscloud.sh:53: response=$(curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" \ diff --git a/.audit/mad-audit-http.log b/.audit/mad-audit-http.log index 764aec6..e912460 100755 --- a/.audit/mad-audit-http.log +++ b/.audit/mad-audit-http.log @@ -1,4 +1,4 @@ -DateTime: 24.12.26 +DateTime: 25.01.04 Files: ./hosts/1fichier.sh @@ -150,7 +150,7 @@ _________________________________________________________________________ ./hosts/9saves.sh:141: --data "$form_data" "https://9saves.com/") ./hosts/anonfile.sh:230: if grep -Eqi 'img src="https://anonfile.de/captchas/' <<< "$response" ; then ./hosts/anonfile.sh:414: if grep -Eqi 'https://.*dashfile.net.*$)' <<< "$response") +./hosts/dashfile.sh:367: if grep -Eqi 'https://.*dashfile.net.*$)' <<< "$response") ./hosts/dataupload.sh:133: post_action="https://dataupload.net/" ./hosts/dataupload.sh:210: if ! grep -Eqi "location: https://dataupload.net/d/" <<< "$response"; then ./hosts/dosya.sh:168: if grep -Eqi 'https://dosyaupload.com' <<< "$dos_url" ; then @@ -179,10 +179,11 @@ _________________________________________________________________________ ./hosts/downloadgg.sh:268: -H "Origin: https://download.gg" \ ./hosts/downloadgg.sh:278: -H "Origin: https://download.gg" \ ./hosts/downloadgg.sh:297: -H "Origin: https://download.gg" \ +./hosts/eternalhosting.sh:36: if grep -Eqi '\.onion' <<< "$pUrlMod" && grep -Eqi 'https://' <<< "$pUrlMod" ; then ./hosts/examples/up_example.sh:105: local ar_HUP[0]='https://oshi.at' -./hosts/fileblade.sh:298: if ! grep -Eqi '^${url/https:/http:}.*>#& #OK# ${filename}>g" "${InputFile}" #http (if changed) -./mad.sh:667: sed -i -e "s>^direct=${url/https:/http:}.*>#& #OK# ${filename}>g" "${InputFile}" #direct url https -./mad.sh:670: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:672: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:693: sed -i -e "s>^${url/https:/http:}.*>#& #OK# ${filename} (File exists)>g" "${InputFile}" #http (if changed) -./mad.sh:695: sed -i -e "s>^direct=${url/https:/http:}.*>#& #OK# ${filename} (File exists)>g" "${InputFile}" #direct url https -./mad.sh:698: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:700: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:721: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${message}>g" "${InputFile}" #http (if changed) -./mad.sh:723: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${message}>g" "${InputFile}" #direct url https -./mad.sh:726: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:728: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:750: sed -i -e "s>^${url/https:/http:}.*>#& #FAIL# $message>g" "${InputFile}" #http (if changed) -./mad.sh:752: sed -i -e "s>^direct=${url/https:/http:}.*>#& #FAIL# $message>g" "${InputFile}" #direct url https -./mad.sh:755: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:757: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:778: sed -i -e "s>^${url/https:/http:}.*>#& #REMOVED#>g" "${InputFile}" #http (if changed) -./mad.sh:780: sed -i -e "s>^direct=${url/https:/http:}.*>#& #REMOVED#>g" "${InputFile}" #direct url https -./mad.sh:783: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:785: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:811: sed -i -e "s>^${url/https:/http:}.*>${url}|${newfilename}>g" "${InputFile}" #http (if changed) -./mad.sh:813: sed -i -e "s>^direct=${url/https:/http:}.*>direct=${url}|${newfilename}>g" "${InputFile}" #direct url https -./mad.sh:833: sed -i -e "s%^${url/https:/http:}.*%${newurl//[[:space:]]/$'\\\n'}%g" "${InputFile}" #http (if changed) -./mad.sh:854: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (dropped or bad size)>g" "${InputFile}" #http (if changed) -./mad.sh:856: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (dropped or bad size)>g" "${InputFile}" #direct url https -./mad.sh:859: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:861: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:877: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# (No CDN found)>g" "${InputFile}" #http (if changed) -./mad.sh:879: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# (No CDN found)>g" "${InputFile}" #direct url https -./mad.sh:882: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:884: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:903: sed -i -e "s>^${url/https:/http:}.*>#& #PASSWORD#>g" "${InputFile}" #http (if changed) -./mad.sh:905: sed -i -e "s>^direct=${url/https:/http:}.*>#& #PASSWORD#>g" "${InputFile}" #direct url https -./mad.sh:908: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:910: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:930: sed -i -e "s>^${url/https:/http:}.*>#& #BAD-URL#${message}>g" "${InputFile}" #http (if changed) -./mad.sh:932: sed -i -e "s>^direct=${url/https:/http:}.*>#& #BAD-URL#${message}>g" "${InputFile}" #direct url https -./mad.sh:935: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:937: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:955: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (No filename or size. Try again later)>g" "${InputFile}" #http (if changed) -./mad.sh:957: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (No filename or size. Try again later)>g" "${InputFile}" #direct url https -./mad.sh:960: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:962: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:981: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (File exists and is too large--retry or delete from downloads)>g" "${InputFile}" #http (if changed) -./mad.sh:983: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (File exists and is too large--retry or delete from downloads)>g" "${InputFile}" #direct url https -./mad.sh:986: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) -./mad.sh:988: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https -./mad.sh:1406: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest) -./mad.sh:1423: download_url='https://github.com/lwthiker/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz' -./mad.sh:1529: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest) -./mad.sh:1546: download_url='https://github.com/lexiforest/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz' -./mad.sh:1809: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" -./mad.sh:1837: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" -./mad.sh:1859: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" -./mad.sh:3175: if grep -Eqi '.onion' <<< "$download_url" && grep -Eqi 'https://' <<< "$download_url" ; then -./mad.sh:3688:arg2="$2" # auto, filelist, -./mad.sh:3785: echo -e " - http://oshi.at/abcd/origAABB.rar|My specified file.part1.rar" -./mad.sh:3787: echo -e " - direct=http://pomf2.lain.la/f/abcd00zz.7z" -./mad.sh:3789: echo -e ' - ie. direct=http://somehost.onion/abcD|filename.part1.rar' -./mad.sh:4008: if [[ ${remote_url} =~ ^http: ]] ; then -./mad.sh:4009: remote_url=${remote_url/http:/https:} -./mad.sh:4030: if [[ ${remote_url} =~ ^http: ]] ; then -./mad.sh:4031: remote_url=${remote_url/http:/https:} -./mad.sh:4397: if [[ ${remote_url} =~ ^http: ]] ; then -./mad.sh:4398: remote_url=${remote_url/http:/https:} -./mad.sh:4456: if [[ ${remote_url} =~ ^http: ]] ; then -./mad.sh:4457: remote_url=${remote_url/http:/https:} -./mad.sh:4482: if [[ ${remote_url} =~ ^http: ]] ; then -./mad.sh:4483: remote_url=${remote_url/http:/https:} +./mad.sh:672: sed -i -e "s>^${url/https:/http:}.*>#& #OK# ${filename}>g" "${InputFile}" #http (if changed) +./mad.sh:674: sed -i -e "s>^direct=${url/https:/http:}.*>#& #OK# ${filename}>g" "${InputFile}" #direct url https +./mad.sh:677: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:679: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:700: sed -i -e "s>^${url/https:/http:}.*>#& #OK# ${filename} (File exists)>g" "${InputFile}" #http (if changed) +./mad.sh:702: sed -i -e "s>^direct=${url/https:/http:}.*>#& #OK# ${filename} (File exists)>g" "${InputFile}" #direct url https +./mad.sh:705: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:707: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:728: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${message}>g" "${InputFile}" #http (if changed) +./mad.sh:730: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${message}>g" "${InputFile}" #direct url https +./mad.sh:733: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:735: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:757: sed -i -e "s>^${url/https:/http:}.*>#& #FAIL# $message>g" "${InputFile}" #http (if changed) +./mad.sh:759: sed -i -e "s>^direct=${url/https:/http:}.*>#& #FAIL# $message>g" "${InputFile}" #direct url https +./mad.sh:762: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:764: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:785: sed -i -e "s>^${url/https:/http:}.*>#& #REMOVED#>g" "${InputFile}" #http (if changed) +./mad.sh:787: sed -i -e "s>^direct=${url/https:/http:}.*>#& #REMOVED#>g" "${InputFile}" #direct url https +./mad.sh:790: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:792: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:818: sed -i -e "s>^${url/https:/http:}.*>${url}|${newfilename}>g" "${InputFile}" #http (if changed) +./mad.sh:820: sed -i -e "s>^direct=${url/https:/http:}.*>direct=${url}|${newfilename}>g" "${InputFile}" #direct url https +./mad.sh:840: sed -i -e "s%^${url/https:/http:}.*%${newurl//[[:space:]]/$'\\\n'}%g" "${InputFile}" #http (if changed) +./mad.sh:861: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (dropped or bad size)>g" "${InputFile}" #http (if changed) +./mad.sh:863: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (dropped or bad size)>g" "${InputFile}" #direct url https +./mad.sh:866: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:868: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:884: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# (No CDN found)>g" "${InputFile}" #http (if changed) +./mad.sh:886: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# (No CDN found)>g" "${InputFile}" #direct url https +./mad.sh:889: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:891: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:910: sed -i -e "s>^${url/https:/http:}.*>#& #PASSWORD#>g" "${InputFile}" #http (if changed) +./mad.sh:912: sed -i -e "s>^direct=${url/https:/http:}.*>#& #PASSWORD#>g" "${InputFile}" #direct url https +./mad.sh:915: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:917: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:937: sed -i -e "s>^${url/https:/http:}.*>#& #BAD-URL#${message}>g" "${InputFile}" #http (if changed) +./mad.sh:939: sed -i -e "s>^direct=${url/https:/http:}.*>#& #BAD-URL#${message}>g" "${InputFile}" #direct url https +./mad.sh:942: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:944: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:962: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (No filename or size. Try again later)>g" "${InputFile}" #http (if changed) +./mad.sh:964: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (No filename or size. Try again later)>g" "${InputFile}" #direct url https +./mad.sh:967: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:969: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:988: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (File exists and is too large--retry or delete from downloads)>g" "${InputFile}" #http (if changed) +./mad.sh:990: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (File exists and is too large--retry or delete from downloads)>g" "${InputFile}" #direct url https +./mad.sh:993: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed) +./mad.sh:995: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https +./mad.sh:1413: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest) +./mad.sh:1430: download_url='https://github.com/lwthiker/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz' +./mad.sh:1536: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest) +./mad.sh:1553: download_url='https://github.com/lexiforest/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz' +./mad.sh:1816: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" +./mad.sh:1844: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" +./mad.sh:1866: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" +./mad.sh:3182: if grep -Eqi '.onion' <<< "$download_url" && grep -Eqi 'https://' <<< "$download_url" ; then +./mad.sh:3695:arg2="$2" # auto, filelist, +./mad.sh:3792: echo -e " - http://oshi.at/abcd/origAABB.rar|My specified file.part1.rar" +./mad.sh:3794: echo -e " - direct=http://pomf2.lain.la/f/abcd00zz.7z" +./mad.sh:3796: echo -e ' - ie. direct=http://somehost.onion/abcD|filename.part1.rar' +./mad.sh:4015: if [[ ${remote_url} =~ ^http: ]] ; then +./mad.sh:4016: remote_url=${remote_url/http:/https:} +./mad.sh:4037: if [[ ${remote_url} =~ ^http: ]] ; then +./mad.sh:4038: remote_url=${remote_url/http:/https:} +./mad.sh:4404: if [[ ${remote_url} =~ ^http: ]] ; then +./mad.sh:4405: remote_url=${remote_url/http:/https:} +./mad.sh:4463: if [[ ${remote_url} =~ ^http: ]] ; then +./mad.sh:4464: remote_url=${remote_url/http:/https:} +./mad.sh:4489: if [[ ${remote_url} =~ ^http: ]] ; then +./mad.sh:4490: remote_url=${remote_url/http:/https:} ./plugins/pjscloud.sh:51: "https://PhantomJScloud.com/api/browser/v2/$RandomPjsKey/" & sleep 8s; kill -HUP $! 2>/dev/null) ./plugins/pjscloud.sh:59: "https://PhantomJScloud.com/api/browser/v2/$RandomPjsKey/" & sleep 8s; kill -HUP $! 2>/dev/null) diff --git a/.audit/mad-audit-tor_curl-details.log b/.audit/mad-audit-tor_curl-details.log index dfcbe0d..650249e 100755 --- a/.audit/mad-audit-tor_curl-details.log +++ b/.audit/mad-audit-tor_curl-details.log @@ -1,4 +1,4 @@ -DateTime: 24.12.26 +DateTime: 25.01.04 Files: ./hosts/1fichier.sh @@ -326,54 +326,54 @@ _________________________________________________________________________ ./hosts/anonfile.sh:349: printf "\\n" ./hosts/anonfile.sh:350: echo -e "${RED}| Failed to extract download link [3].${NC}" -- -./hosts/anonfile.sh:446: file_header=$(tor_curl_request -i -s --head \ -./hosts/anonfile.sh:447: --referer "${fixed_url}" \ -./hosts/anonfile.sh:448: "$download_url") -./hosts/anonfile.sh:449: if [ "${DebugAllEnabled}" == "true" ] ; then -./hosts/anonfile.sh:450: debugHtml "${remote_url##*/}" "anon_head$j" "download_url: ${download_url}"$'\n'"${file_header}" -./hosts/anonfile.sh:451: fi -./hosts/anonfile.sh:452: if [[ -z $file_header ]] ; then -./hosts/anonfile.sh:453: if [ $j == $maxfetchretries ] ; then -./hosts/anonfile.sh:454: rm -f "${anon_cookie_jar}"; -./hosts/anonfile.sh:455: printf "\\n" -./hosts/anonfile.sh:456: echo -e "${RED}| Failed to extract file info${NC}" +./hosts/anonfile.sh:451: file_header=$(tor_curl_request -i -s --head \ +./hosts/anonfile.sh:452: --referer "${fixed_url}" \ +./hosts/anonfile.sh:453: "$download_url") +./hosts/anonfile.sh:454: if [ "${DebugAllEnabled}" == "true" ] ; then +./hosts/anonfile.sh:455: debugHtml "${remote_url##*/}" "anon_head$j" "download_url: ${download_url}"$'\n'"${file_header}" +./hosts/anonfile.sh:456: fi +./hosts/anonfile.sh:457: if [[ -z $file_header ]] ; then +./hosts/anonfile.sh:458: if [ $j == $maxfetchretries ] ; then +./hosts/anonfile.sh:459: rm -f "${anon_cookie_jar}"; +./hosts/anonfile.sh:460: printf "\\n" +./hosts/anonfile.sh:461: echo -e "${RED}| Failed to extract file info${NC}" -- -./hosts/anonfile.sh:551: tor_curl_request --insecure \ -./hosts/anonfile.sh:552: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ -./hosts/anonfile.sh:553: -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \ -./hosts/anonfile.sh:554: --referer "${fixed_url}" \ -./hosts/anonfile.sh:555: "$download_url" --continue-at - --output "$file_path" -./hosts/anonfile.sh:556: else -./hosts/anonfile.sh:557: tor_curl_request --insecure \ -./hosts/anonfile.sh:558: -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \ -./hosts/anonfile.sh:559: --referer "${fixed_url}" \ -./hosts/anonfile.sh:560: "$download_url" --continue-at - --output "$file_path" -./hosts/anonfile.sh:561: fi -./hosts/anonfile.sh:562: else -./hosts/anonfile.sh:563: if [ "${RateMonitorEnabled}" == "true" ]; then -./hosts/anonfile.sh:564: tor_curl_request --insecure \ -./hosts/anonfile.sh:565: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ -./hosts/anonfile.sh:566: -H "User-Agent: $RandomUA" \ -./hosts/anonfile.sh:567: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ -./hosts/anonfile.sh:568: -H "Accept-Language: en-US,en;q=0.5" \ -./hosts/anonfile.sh:569: -H "Accept-Encoding: gzip, deflate, br" \ -./hosts/anonfile.sh:570: -H "Connection: keep-alive" \ -./hosts/anonfile.sh:571: -H "Cookie: lng=eng" \ -./hosts/anonfile.sh:572: -H "Upgrade-Insecure-Requests: 1" \ -./hosts/anonfile.sh:573: -H "Sec-Fetch-Dest: document" \ -./hosts/anonfile.sh:574: -H "Sec-Fetch-Mode: navigate" \ +./hosts/anonfile.sh:559: tor_curl_request --insecure \ +./hosts/anonfile.sh:560: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ +./hosts/anonfile.sh:561: -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \ +./hosts/anonfile.sh:562: --referer "${fixed_url}" \ +./hosts/anonfile.sh:563: "$download_url" --output "$file_path" +./hosts/anonfile.sh:564: else +./hosts/anonfile.sh:565: tor_curl_request --insecure \ +./hosts/anonfile.sh:566: -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \ +./hosts/anonfile.sh:567: --referer "${fixed_url}" \ +./hosts/anonfile.sh:568: "$download_url" --output "$file_path" +./hosts/anonfile.sh:569: fi +./hosts/anonfile.sh:570: else +./hosts/anonfile.sh:571: if [ "${RateMonitorEnabled}" == "true" ]; then +./hosts/anonfile.sh:572: tor_curl_request --insecure \ +./hosts/anonfile.sh:573: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ +./hosts/anonfile.sh:574: -H "User-Agent: $RandomUA" \ +./hosts/anonfile.sh:575: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ +./hosts/anonfile.sh:576: -H "Accept-Language: en-US,en;q=0.5" \ +./hosts/anonfile.sh:577: -H "Accept-Encoding: gzip, deflate, br" \ +./hosts/anonfile.sh:578: -H "Connection: keep-alive" \ +./hosts/anonfile.sh:579: -H "Cookie: lng=eng" \ +./hosts/anonfile.sh:580: -H "Upgrade-Insecure-Requests: 1" \ +./hosts/anonfile.sh:581: -H "Sec-Fetch-Dest: document" \ +./hosts/anonfile.sh:582: -H "Sec-Fetch-Mode: navigate" \ -- -./hosts/anonfile.sh:581: tor_curl_request --insecure \ -./hosts/anonfile.sh:582: -H "User-Agent: $RandomUA" \ -./hosts/anonfile.sh:583: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ -./hosts/anonfile.sh:584: -H "Accept-Language: en-US,en;q=0.5" \ -./hosts/anonfile.sh:585: -H "Accept-Encoding: gzip, deflate, br" \ -./hosts/anonfile.sh:586: -H "Connection: keep-alive" \ -./hosts/anonfile.sh:587: -H "Cookie: lng=eng" \ -./hosts/anonfile.sh:588: -H "Upgrade-Insecure-Requests: 1" \ -./hosts/anonfile.sh:589: -H "Sec-Fetch-Dest: document" \ -./hosts/anonfile.sh:590: -H "Sec-Fetch-Mode: navigate" \ -./hosts/anonfile.sh:591: -H "Sec-Fetch-Site: same-origin" \ +./hosts/anonfile.sh:589: tor_curl_request --insecure \ +./hosts/anonfile.sh:590: -H "User-Agent: $RandomUA" \ +./hosts/anonfile.sh:591: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ +./hosts/anonfile.sh:592: -H "Accept-Language: en-US,en;q=0.5" \ +./hosts/anonfile.sh:593: -H "Accept-Encoding: gzip, deflate, br" \ +./hosts/anonfile.sh:594: -H "Connection: keep-alive" \ +./hosts/anonfile.sh:595: -H "Cookie: lng=eng" \ +./hosts/anonfile.sh:596: -H "Upgrade-Insecure-Requests: 1" \ +./hosts/anonfile.sh:597: -H "Sec-Fetch-Dest: document" \ +./hosts/anonfile.sh:598: -H "Sec-Fetch-Mode: navigate" \ +./hosts/anonfile.sh:599: -H "Sec-Fetch-Site: same-origin" \ -- ./hosts/anonsharing.sh:91: response=$(tor_curl_request --insecure -i -s \ ./hosts/anonsharing.sh:92: -b "${ansh_cookie_jar}" -c "${ansh_cookie_jar}" \ @@ -829,76 +829,76 @@ _________________________________________________________________________ ./hosts/dashfile.sh:99: printf "\\n" ./hosts/dashfile.sh:100: echo -e "${RED}| Failed to extract download link [1]${NC}" -- -./hosts/dashfile.sh:169: response=$(tor_curl_request --insecure -L -s -X POST \ -./hosts/dashfile.sh:170: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ -./hosts/dashfile.sh:171: --data "$form_data" "$remote_url") -./hosts/dashfile.sh:172: if [ "${DebugAllEnabled}" == "true" ] ; then -./hosts/dashfile.sh:173: debugHtml "${remote_url##*/}" "dash_post1_$i" "url: ${remote_url}"$'\n'"form_data: ${form_data}"$'\n'"${response}" -./hosts/dashfile.sh:174: fi -./hosts/dashfile.sh:175: if [[ -z $response ]] ; then -./hosts/dashfile.sh:176: if [ $i == $maxfetchretries ] ; then -./hosts/dashfile.sh:177: rm -f "${dash_cookie_jar}"; -./hosts/dashfile.sh:178: printf "\\n" -./hosts/dashfile.sh:179: echo -e "${RED}| Failed to extract download link [4]${NC}" +./hosts/dashfile.sh:177: response=$(tor_curl_request --insecure -L -s -X POST \ +./hosts/dashfile.sh:178: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ +./hosts/dashfile.sh:179: --data "$form_data" "$remote_url") +./hosts/dashfile.sh:180: if [ "${DebugAllEnabled}" == "true" ] ; then +./hosts/dashfile.sh:181: debugHtml "${remote_url##*/}" "dash_post1_$i" "url: ${remote_url}"$'\n'"form_data: ${form_data}"$'\n'"${response}" +./hosts/dashfile.sh:182: fi +./hosts/dashfile.sh:183: if [[ -z $response ]] ; then +./hosts/dashfile.sh:184: if [ $i == $maxfetchretries ] ; then +./hosts/dashfile.sh:185: rm -f "${dash_cookie_jar}"; +./hosts/dashfile.sh:186: printf "\\n" +./hosts/dashfile.sh:187: echo -e "${RED}| Failed to extract download link [4]${NC}" -- -./hosts/dashfile.sh:297: response=$(tor_curl_request --insecure -L -s -X POST \ -./hosts/dashfile.sh:298: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ -./hosts/dashfile.sh:299: --data "$form_data" "$remote_url") -./hosts/dashfile.sh:300: if [ "${DebugAllEnabled}" == "true" ] ; then -./hosts/dashfile.sh:301: debugHtml "${remote_url##*/}" "dash_post2_$i" "url: ${remote_url}"$'\n'"form_data: ${form_data}"$'\n'"${response}" -./hosts/dashfile.sh:302: fi -./hosts/dashfile.sh:303: if [[ -z $response ]] ; then -./hosts/dashfile.sh:304: if [ $i == $maxfetchretries ] ; then -./hosts/dashfile.sh:305: rm -f "${dash_cookie_jar}"; -./hosts/dashfile.sh:306: printf "\\n" -./hosts/dashfile.sh:307: echo -e "${RED}| Failed to extract download link [7]${NC}" +./hosts/dashfile.sh:308: response=$(tor_curl_request --insecure -L -s -X POST \ +./hosts/dashfile.sh:309: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ +./hosts/dashfile.sh:310: --data "$form_data" "$remote_url") +./hosts/dashfile.sh:311: if [ "${DebugAllEnabled}" == "true" ] ; then +./hosts/dashfile.sh:312: debugHtml "${remote_url##*/}" "dash_post2_$i" "url: ${remote_url}"$'\n'"form_data: ${form_data}"$'\n'"${response}" +./hosts/dashfile.sh:313: fi +./hosts/dashfile.sh:314: if [[ -z $response ]] ; then +./hosts/dashfile.sh:315: if [ $i == $maxfetchretries ] ; then +./hosts/dashfile.sh:316: rm -f "${dash_cookie_jar}"; +./hosts/dashfile.sh:317: printf "\\n" +./hosts/dashfile.sh:318: echo -e "${RED}| Failed to extract download link [7]${NC}" -- -./hosts/dashfile.sh:386: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url") -./hosts/dashfile.sh:387: if [ "${DebugAllEnabled}" == "true" ] ; then -./hosts/dashfile.sh:388: debugHtml "${remote_url##*/}" "dash_head$j" "download_url: ${download_url}"$'\n'"${file_header}" -./hosts/dashfile.sh:389: fi -./hosts/dashfile.sh:390: if [[ -z $file_header ]] ; then -./hosts/dashfile.sh:391: if [ $j == $maxfetchretries ] ; then -./hosts/dashfile.sh:392: rm -f "${dash_cookie_jar}"; -./hosts/dashfile.sh:393: printf "\\n" -./hosts/dashfile.sh:394: echo -e "${RED}| Failed to extract file info${NC}" -./hosts/dashfile.sh:395: warnAndRetryUnknownError=true -./hosts/dashfile.sh:396: if [ "${finalAttempt}" == "true" ] ; then +./hosts/dashfile.sh:397: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url") +./hosts/dashfile.sh:398: if [ "${DebugAllEnabled}" == "true" ] ; then +./hosts/dashfile.sh:399: debugHtml "${remote_url##*/}" "dash_head$j" "download_url: ${download_url}"$'\n'"${file_header}" +./hosts/dashfile.sh:400: fi +./hosts/dashfile.sh:401: if [[ -z $file_header ]] ; then +./hosts/dashfile.sh:402: if [ $j == $maxfetchretries ] ; then +./hosts/dashfile.sh:403: rm -f "${dash_cookie_jar}"; +./hosts/dashfile.sh:404: printf "\\n" +./hosts/dashfile.sh:405: echo -e "${RED}| Failed to extract file info${NC}" +./hosts/dashfile.sh:406: warnAndRetryUnknownError=true +./hosts/dashfile.sh:407: if [ "${finalAttempt}" == "true" ] ; then -- -./hosts/dashfile.sh:486: tor_curl_request --insecure \ -./hosts/dashfile.sh:487: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ -./hosts/dashfile.sh:488: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ -./hosts/dashfile.sh:489: "$download_url" --continue-at - --output "$file_path" -./hosts/dashfile.sh:490: else -./hosts/dashfile.sh:491: tor_curl_request --insecure \ -./hosts/dashfile.sh:492: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ -./hosts/dashfile.sh:493: "$download_url" --continue-at - --output "$file_path" -./hosts/dashfile.sh:494: fi -./hosts/dashfile.sh:495: else -./hosts/dashfile.sh:496: if [ "${RateMonitorEnabled}" == "true" ]; then ./hosts/dashfile.sh:497: tor_curl_request --insecure \ ./hosts/dashfile.sh:498: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ ./hosts/dashfile.sh:499: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ -./hosts/dashfile.sh:500: -H "User-Agent: $RandomUA" \ -./hosts/dashfile.sh:501: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ -./hosts/dashfile.sh:502: -H "Accept-Language: en-US,en;q=0.5" \ -./hosts/dashfile.sh:503: -H "Accept-Encoding: gzip, deflate, br" \ -./hosts/dashfile.sh:504: -H "Connection: keep-alive" \ -./hosts/dashfile.sh:505: -H "Cookie: lng=eng" \ -./hosts/dashfile.sh:506: -H "Upgrade-Insecure-Requests: 1" \ -./hosts/dashfile.sh:507: -H "Sec-Fetch-Dest: document" \ +./hosts/dashfile.sh:500: "$download_url" --continue-at - --output "$file_path" +./hosts/dashfile.sh:501: else +./hosts/dashfile.sh:502: tor_curl_request --insecure \ +./hosts/dashfile.sh:503: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ +./hosts/dashfile.sh:504: "$download_url" --continue-at - --output "$file_path" +./hosts/dashfile.sh:505: fi +./hosts/dashfile.sh:506: else +./hosts/dashfile.sh:507: if [ "${RateMonitorEnabled}" == "true" ]; then +./hosts/dashfile.sh:508: tor_curl_request --insecure \ +./hosts/dashfile.sh:509: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ +./hosts/dashfile.sh:510: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ +./hosts/dashfile.sh:511: -H "User-Agent: $RandomUA" \ +./hosts/dashfile.sh:512: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ +./hosts/dashfile.sh:513: -H "Accept-Language: en-US,en;q=0.5" \ +./hosts/dashfile.sh:514: -H "Accept-Encoding: gzip, deflate, br" \ +./hosts/dashfile.sh:515: -H "Connection: keep-alive" \ +./hosts/dashfile.sh:516: -H "Cookie: lng=eng" \ +./hosts/dashfile.sh:517: -H "Upgrade-Insecure-Requests: 1" \ +./hosts/dashfile.sh:518: -H "Sec-Fetch-Dest: document" \ -- -./hosts/dashfile.sh:513: tor_curl_request --insecure \ -./hosts/dashfile.sh:514: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ -./hosts/dashfile.sh:515: -H "User-Agent: $RandomUA" \ -./hosts/dashfile.sh:516: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ -./hosts/dashfile.sh:517: -H "Accept-Language: en-US,en;q=0.5" \ -./hosts/dashfile.sh:518: -H "Accept-Encoding: gzip, deflate, br" \ -./hosts/dashfile.sh:519: -H "Connection: keep-alive" \ -./hosts/dashfile.sh:520: -H "Cookie: lng=eng" \ -./hosts/dashfile.sh:521: -H "Upgrade-Insecure-Requests: 1" \ -./hosts/dashfile.sh:522: -H "Sec-Fetch-Dest: document" \ -./hosts/dashfile.sh:523: -H "Sec-Fetch-Mode: navigate" \ +./hosts/dashfile.sh:524: tor_curl_request --insecure \ +./hosts/dashfile.sh:525: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \ +./hosts/dashfile.sh:526: -H "User-Agent: $RandomUA" \ +./hosts/dashfile.sh:527: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ +./hosts/dashfile.sh:528: -H "Accept-Language: en-US,en;q=0.5" \ +./hosts/dashfile.sh:529: -H "Accept-Encoding: gzip, deflate, br" \ +./hosts/dashfile.sh:530: -H "Connection: keep-alive" \ +./hosts/dashfile.sh:531: -H "Cookie: lng=eng" \ +./hosts/dashfile.sh:532: -H "Upgrade-Insecure-Requests: 1" \ +./hosts/dashfile.sh:533: -H "Sec-Fetch-Dest: document" \ +./hosts/dashfile.sh:534: -H "Sec-Fetch-Mode: navigate" \ -- ./hosts/dataupload.sh:90: response=$(tor_curl_request --insecure -L -s -b "${dup_cookie_jar}" -c "${dup_cookie_jar}" "$remote_url") ./hosts/dataupload.sh:91: if [ "${DebugAllEnabled}" == "true" ] ; then @@ -1147,74 +1147,74 @@ _________________________________________________________________________ ./hosts/fileblade.sh:99: warnAndRetryUnknownError=true ./hosts/fileblade.sh:100: if [ "${finalAttempt}" == "true" ] ; then -- -./hosts/fileblade.sh:167: response=$(tor_curl_request --insecure -L -s -X POST \ -./hosts/fileblade.sh:168: -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \ -./hosts/fileblade.sh:169: --data "$form_data" "$post_action") -./hosts/fileblade.sh:170: if [ "${DebugAllEnabled}" == "true" ] ; then -./hosts/fileblade.sh:171: debugHtml "${remote_url##*/}" "fb_post" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}" -./hosts/fileblade.sh:172: fi -./hosts/fileblade.sh:173: if [[ -z $response ]] ; then -./hosts/fileblade.sh:174: if [ $i == $maxfetchretries ] ; then -./hosts/fileblade.sh:175: rm -f "${fb_cookie_jar}"; -./hosts/fileblade.sh:176: printf "\\n" -./hosts/fileblade.sh:177: echo -e "${RED}| Failed to extract download link [3]${NC}" +./hosts/fileblade.sh:164: response=$(tor_curl_request --insecure -L -s -X POST \ +./hosts/fileblade.sh:165: -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \ +./hosts/fileblade.sh:166: --data "$form_data" "$post_action") +./hosts/fileblade.sh:167: if [ "${DebugAllEnabled}" == "true" ] ; then +./hosts/fileblade.sh:168: debugHtml "${remote_url##*/}" "fb_post(1)" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}" +./hosts/fileblade.sh:169: fi +./hosts/fileblade.sh:170: if [[ -z $response ]] ; then +./hosts/fileblade.sh:171: if [ $i == $maxfetchretries ] ; then +./hosts/fileblade.sh:172: rm -f "${fb_cookie_jar}"; +./hosts/fileblade.sh:173: printf "\\n" +./hosts/fileblade.sh:174: echo -e "${RED}| Failed to extract download link [3]${NC}" -- -./hosts/fileblade.sh:254: response=$(tor_curl_request --insecure -L -s -X POST \ -./hosts/fileblade.sh:255: -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \ -./hosts/fileblade.sh:256: --data "$form_data" "$post_action") -./hosts/fileblade.sh:257: if [ "${DebugAllEnabled}" == "true" ] ; then -./hosts/fileblade.sh:258: debugHtml "${remote_url##*/}" "fb_post2" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}" -./hosts/fileblade.sh:259: fi -./hosts/fileblade.sh:260: if [[ -z $response ]] ; then -./hosts/fileblade.sh:261: if [ $i == $maxfetchretries ] ; then -./hosts/fileblade.sh:262: rm -f "${fb_cookie_jar}"; -./hosts/fileblade.sh:263: printf "\\n" -./hosts/fileblade.sh:264: echo -e "${RED}| Failed to extract download link [4].${NC}" +./hosts/fileblade.sh:266: response=$(tor_curl_request --insecure -L -s -X POST \ +./hosts/fileblade.sh:267: -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \ +./hosts/fileblade.sh:268: --data "$form_data" "$post_action") +./hosts/fileblade.sh:269: if [ "${DebugAllEnabled}" == "true" ] ; then +./hosts/fileblade.sh:270: debugHtml "${remote_url##*/}" "fb_post(2)" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}" +./hosts/fileblade.sh:271: fi +./hosts/fileblade.sh:272: if [[ -z $response ]] ; then +./hosts/fileblade.sh:273: if [ $i == $maxfetchretries ] ; then +./hosts/fileblade.sh:274: rm -f "${fb_cookie_jar}"; +./hosts/fileblade.sh:275: printf "\\n" +./hosts/fileblade.sh:276: echo -e "${RED}| Failed to extract download link [4].${NC}" -- -./hosts/fileblade.sh:323: file_header=$(tor_curl_request --insecure -L --head -s "$download_url") -./hosts/fileblade.sh:324: if [ "${DebugAllEnabled}" == "true" ] ; then -./hosts/fileblade.sh:325: debugHtml "${remote_url##*/}" "fb_head$j" "download_url: ${download_url}"$'\n'"${file_header}" -./hosts/fileblade.sh:326: fi -./hosts/fileblade.sh:327: if [[ -z $file_header ]] ; then -./hosts/fileblade.sh:328: if [ $j == $maxfetchretries ] ; then -./hosts/fileblade.sh:329: rm -f "${fb_cookie_jar}"; -./hosts/fileblade.sh:330: printf "\\n" -./hosts/fileblade.sh:331: echo -e "${RED}| Failed to extract file info [1]${NC}" -./hosts/fileblade.sh:332: warnAndRetryUnknownError=true -./hosts/fileblade.sh:333: if [ "${finalAttempt}" == "true" ] ; then +./hosts/fileblade.sh:335: file_header=$(tor_curl_request --insecure -L --head -s "$download_url") +./hosts/fileblade.sh:336: if [ "${DebugAllEnabled}" == "true" ] ; then +./hosts/fileblade.sh:337: debugHtml "${remote_url##*/}" "fb_head$j" "download_url: ${download_url}"$'\n'"${file_header}" +./hosts/fileblade.sh:338: fi +./hosts/fileblade.sh:339: if [[ -z $file_header ]] ; then +./hosts/fileblade.sh:340: if [ $j == $maxfetchretries ] ; then +./hosts/fileblade.sh:341: rm -f "${fb_cookie_jar}"; +./hosts/fileblade.sh:342: printf "\\n" +./hosts/fileblade.sh:343: echo -e "${RED}| Failed to extract file info [1]${NC}" +./hosts/fileblade.sh:344: warnAndRetryUnknownError=true +./hosts/fileblade.sh:345: if [ "${finalAttempt}" == "true" ] ; then -- -./hosts/fileblade.sh:440: tor_curl_request --insecure -L \ -./hosts/fileblade.sh:441: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ -./hosts/fileblade.sh:442: "$download_url" --continue-at - --output "$file_path" -./hosts/fileblade.sh:443: else -./hosts/fileblade.sh:444: tor_curl_request --insecure -L \ -./hosts/fileblade.sh:445: "$download_url" --continue-at - --output "$file_path" -./hosts/fileblade.sh:446: fi -./hosts/fileblade.sh:447: else -./hosts/fileblade.sh:448: if [ "${RateMonitorEnabled}" == "true" ]; then -./hosts/fileblade.sh:449: tor_curl_request --insecure \ -./hosts/fileblade.sh:450: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ -./hosts/fileblade.sh:451: -H "User-Agent: $RandomUA" \ -./hosts/fileblade.sh:452: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ -./hosts/fileblade.sh:453: -H "Accept-Language: en-US,en;q=0.5" \ -./hosts/fileblade.sh:454: -H "Accept-Encoding: gzip, deflate, br" \ -./hosts/fileblade.sh:455: -H "Connection: keep-alive" \ -./hosts/fileblade.sh:456: -H "Cookie: lng=eng" \ -./hosts/fileblade.sh:457: -H "Upgrade-Insecure-Requests: 1" \ -./hosts/fileblade.sh:458: -H "Sec-Fetch-Dest: document" \ -./hosts/fileblade.sh:459: -H "Sec-Fetch-Mode: navigate" \ +./hosts/fileblade.sh:452: tor_curl_request --insecure -L \ +./hosts/fileblade.sh:453: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ +./hosts/fileblade.sh:454: "$download_url" --continue-at - --output "$file_path" +./hosts/fileblade.sh:455: else +./hosts/fileblade.sh:456: tor_curl_request --insecure -L \ +./hosts/fileblade.sh:457: "$download_url" --continue-at - --output "$file_path" +./hosts/fileblade.sh:458: fi +./hosts/fileblade.sh:459: else +./hosts/fileblade.sh:460: if [ "${RateMonitorEnabled}" == "true" ]; then +./hosts/fileblade.sh:461: tor_curl_request --insecure \ +./hosts/fileblade.sh:462: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ +./hosts/fileblade.sh:463: -H "User-Agent: $RandomUA" \ +./hosts/fileblade.sh:464: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ +./hosts/fileblade.sh:465: -H "Accept-Language: en-US,en;q=0.5" \ +./hosts/fileblade.sh:466: -H "Accept-Encoding: gzip, deflate, br" \ +./hosts/fileblade.sh:467: -H "Connection: keep-alive" \ +./hosts/fileblade.sh:468: -H "Cookie: lng=eng" \ +./hosts/fileblade.sh:469: -H "Upgrade-Insecure-Requests: 1" \ +./hosts/fileblade.sh:470: -H "Sec-Fetch-Dest: document" \ +./hosts/fileblade.sh:471: -H "Sec-Fetch-Mode: navigate" \ -- -./hosts/fileblade.sh:464: tor_curl_request --insecure \ -./hosts/fileblade.sh:465: -H "User-Agent: $RandomUA" \ -./hosts/fileblade.sh:466: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ -./hosts/fileblade.sh:467: -H "Accept-Language: en-US,en;q=0.5" \ -./hosts/fileblade.sh:468: -H "Accept-Encoding: gzip, deflate, br" \ -./hosts/fileblade.sh:469: -H "Connection: keep-alive" \ -./hosts/fileblade.sh:470: -H "Cookie: lng=eng" \ -./hosts/fileblade.sh:471: -H "Upgrade-Insecure-Requests: 1" \ -./hosts/fileblade.sh:472: -H "Sec-Fetch-Dest: document" \ -./hosts/fileblade.sh:473: -H "Sec-Fetch-Mode: navigate" \ -./hosts/fileblade.sh:474: -H "Sec-Fetch-Site: same-origin" \ +./hosts/fileblade.sh:476: tor_curl_request --insecure \ +./hosts/fileblade.sh:477: -H "User-Agent: $RandomUA" \ +./hosts/fileblade.sh:478: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ +./hosts/fileblade.sh:479: -H "Accept-Language: en-US,en;q=0.5" \ +./hosts/fileblade.sh:480: -H "Accept-Encoding: gzip, deflate, br" \ +./hosts/fileblade.sh:481: -H "Connection: keep-alive" \ +./hosts/fileblade.sh:482: -H "Cookie: lng=eng" \ +./hosts/fileblade.sh:483: -H "Upgrade-Insecure-Requests: 1" \ +./hosts/fileblade.sh:484: -H "Sec-Fetch-Dest: document" \ +./hosts/fileblade.sh:485: -H "Sec-Fetch-Mode: navigate" \ +./hosts/fileblade.sh:486: -H "Sec-Fetch-Site: same-origin" \ -- ./hosts/fileditch.sh:85: file_header=$(tor_curl_request --insecure --head -L -s "$download_url") ./hosts/fileditch.sh:86: if [ "${DebugAllEnabled}" == "true" ] ; then @@ -1404,41 +1404,41 @@ _________________________________________________________________________ ./hosts/gofile.sh:179: -H "Sec-Fetch-Mode: navigate" \ ./hosts/gofile.sh:180: -H "Sec-Fetch-Site: none" \ -- -./hosts/gofile.sh:250: file_header=$(tor_curl_request --insecure -L --head -s \ -./hosts/gofile.sh:251: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \ -./hosts/gofile.sh:252: -H "User-Agent: $RandomUA" \ -./hosts/gofile.sh:253: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ -./hosts/gofile.sh:254: -H "Accept-Language: en-US,en;q=0.5" \ -./hosts/gofile.sh:255: -H "Accept-Encoding: gzip, deflate, br" \ -./hosts/gofile.sh:256: -H "Connection: keep-alive" \ -./hosts/gofile.sh:257: -H "Upgrade-Insecure-Requests: 1" \ -./hosts/gofile.sh:258: -H "Sec-Fetch-Dest: document" \ -./hosts/gofile.sh:259: -H "Sec-Fetch-Mode: navigate" \ -./hosts/gofile.sh:260: -H "Sec-Fetch-Site: none" \ +./hosts/gofile.sh:258: file_header=$(tor_curl_request --insecure -L --head -s \ +./hosts/gofile.sh:259: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \ +./hosts/gofile.sh:260: -H "User-Agent: $RandomUA" \ +./hosts/gofile.sh:261: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ +./hosts/gofile.sh:262: -H "Accept-Language: en-US,en;q=0.5" \ +./hosts/gofile.sh:263: -H "Accept-Encoding: gzip, deflate, br" \ +./hosts/gofile.sh:264: -H "Connection: keep-alive" \ +./hosts/gofile.sh:265: -H "Upgrade-Insecure-Requests: 1" \ +./hosts/gofile.sh:266: -H "Sec-Fetch-Dest: document" \ +./hosts/gofile.sh:267: -H "Sec-Fetch-Mode: navigate" \ +./hosts/gofile.sh:268: -H "Sec-Fetch-Site: none" \ -- -./hosts/gofile.sh:369: tor_curl_request --insecure -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ -./hosts/gofile.sh:370: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \ -./hosts/gofile.sh:371: -H "User-Agent: $RandomUA" \ -./hosts/gofile.sh:372: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ -./hosts/gofile.sh:373: -H "Accept-Language: en-US,en;q=0.5" \ -./hosts/gofile.sh:374: -H "Accept-Encoding: gzip, deflate, br" \ -./hosts/gofile.sh:375: -H "Connection: keep-alive" \ -./hosts/gofile.sh:376: -H "Upgrade-Insecure-Requests: 1" \ -./hosts/gofile.sh:377: -H "Sec-Fetch-Dest: document" \ -./hosts/gofile.sh:378: -H "Sec-Fetch-Mode: navigate" \ -./hosts/gofile.sh:379: -H "Sec-Fetch-Site: none" \ +./hosts/gofile.sh:377: tor_curl_request --insecure -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ +./hosts/gofile.sh:378: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \ +./hosts/gofile.sh:379: -H "User-Agent: $RandomUA" \ +./hosts/gofile.sh:380: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ +./hosts/gofile.sh:381: -H "Accept-Language: en-US,en;q=0.5" \ +./hosts/gofile.sh:382: -H "Accept-Encoding: gzip, deflate, br" \ +./hosts/gofile.sh:383: -H "Connection: keep-alive" \ +./hosts/gofile.sh:384: -H "Upgrade-Insecure-Requests: 1" \ +./hosts/gofile.sh:385: -H "Sec-Fetch-Dest: document" \ +./hosts/gofile.sh:386: -H "Sec-Fetch-Mode: navigate" \ +./hosts/gofile.sh:387: -H "Sec-Fetch-Site: none" \ -- -./hosts/gofile.sh:383: tor_curl_request --insecure -G \ -./hosts/gofile.sh:384: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \ -./hosts/gofile.sh:385: -H "User-Agent: $RandomUA" \ -./hosts/gofile.sh:386: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ -./hosts/gofile.sh:387: -H "Accept-Language: en-US,en;q=0.5" \ -./hosts/gofile.sh:388: -H "Accept-Encoding: gzip, deflate, br" \ -./hosts/gofile.sh:389: -H "Connection: keep-alive" \ -./hosts/gofile.sh:390: -H "Upgrade-Insecure-Requests: 1" \ -./hosts/gofile.sh:391: -H "Sec-Fetch-Dest: document" \ -./hosts/gofile.sh:392: -H "Sec-Fetch-Mode: navigate" \ -./hosts/gofile.sh:393: -H "Sec-Fetch-Site: none" \ +./hosts/gofile.sh:391: tor_curl_request --insecure -G \ +./hosts/gofile.sh:392: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \ +./hosts/gofile.sh:393: -H "User-Agent: $RandomUA" \ +./hosts/gofile.sh:394: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \ +./hosts/gofile.sh:395: -H "Accept-Language: en-US,en;q=0.5" \ +./hosts/gofile.sh:396: -H "Accept-Encoding: gzip, deflate, br" \ +./hosts/gofile.sh:397: -H "Connection: keep-alive" \ +./hosts/gofile.sh:398: -H "Upgrade-Insecure-Requests: 1" \ +./hosts/gofile.sh:399: -H "Sec-Fetch-Dest: document" \ +./hosts/gofile.sh:400: -H "Sec-Fetch-Mode: navigate" \ +./hosts/gofile.sh:401: -H "Sec-Fetch-Site: none" \ -- ./hosts/hexload.sh:108: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download") ./hosts/hexload.sh:109: if [ "${DebugAllEnabled}" == "true" ] ; then @@ -2511,8 +2511,8 @@ _________________________________________________________________________ ./hosts/up_axfc.sh:190: if [ "${DebugAllEnabled}" == "true" ] ; then ./hosts/up_axfc.sh:191: debugHtml "${filepath##*/}" "${_hostCode}_upload" "post_url: ${PostUrlHost}"$'\n'"${response}" ./hosts/up_axfc.sh:192: fi -./hosts/up_axfc.sh:193: if grep -Eqi 'Axfc Uploader -投稿完了.*キーワード付きURL:.*a href="https://www.axfc.net.*(QueryString無しVer)' <<< "${response}" ; then -./hosts/up_axfc.sh:194: subSearch=$(awk '/Axfc Uploader -投稿完了/,/(QueryString無しVer)/' <<< "$response") +./hosts/up_axfc.sh:193: response_ascii=$(echo "$response" | iconv -c -f UTF-8 -t ASCII//TRANSLIT) +./hosts/up_axfc.sh:194: if [ "${DebugAllEnabled}" == "true" ] ; then -- ./hosts/up_bedrive.sh:102: response=$(tor_curl_upload --insecure -i \ ./hosts/up_bedrive.sh:103: -H "Content-Type: multipart/form-data" \ @@ -3068,235 +3068,235 @@ _________________________________________________________________________ ./hosts/youdbox.sh:287: containsHtml=true ./hosts/youdbox.sh:288: fi -- -./mad.sh:391:tor_curl_request() { -./mad.sh:392: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:393: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" -./mad.sh:394: else -./mad.sh:395: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" -./mad.sh:396: fi -./mad.sh:397:} -./mad.sh:398:tor_curl_request_extended() { -./mad.sh:399: randomtimeout=$((30 + RANDOM % (60 - 30))) -./mad.sh:400: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:401: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@" -./mad.sh:402: else -./mad.sh:403: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@" -./mad.sh:404: fi -./mad.sh:405:} -./mad.sh:406:tor_curl_upload() { +./mad.sh:398:tor_curl_request() { +./mad.sh:399: if [ "${UseTorCurlImpersonate}" == "true" ]; then +./mad.sh:400: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" +./mad.sh:401: else +./mad.sh:402: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" +./mad.sh:403: fi +./mad.sh:404:} +./mad.sh:405:tor_curl_request_extended() { +./mad.sh:406: randomtimeout=$((30 + RANDOM % (60 - 30))) ./mad.sh:407: if [ "${UseTorCurlImpersonate}" == "true" ]; then -./mad.sh:408: if [ "${RateMonitorEnabled}" == "true" ]; then -./mad.sh:409: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval --compressed --globoff "$@" -./mad.sh:410: else -./mad.sh:411: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" -./mad.sh:412: fi -./mad.sh:413: else -./mad.sh:414: if [ "${RateMonitorEnabled}" == "true" ]; then -./mad.sh:415: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@" -./mad.sh:416: else +./mad.sh:408: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@" +./mad.sh:409: else +./mad.sh:410: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@" +./mad.sh:411: fi +./mad.sh:412:} +./mad.sh:413:tor_curl_upload() { +./mad.sh:414: if [ "${UseTorCurlImpersonate}" == "true" ]; then +./mad.sh:415: if [ "${RateMonitorEnabled}" == "true" ]; then +./mad.sh:416: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval --compressed --globoff "$@" +./mad.sh:417: else +./mad.sh:418: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@" +./mad.sh:419: fi +./mad.sh:420: else +./mad.sh:421: if [ "${RateMonitorEnabled}" == "true" ]; then +./mad.sh:422: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@" +./mad.sh:423: else -- -./mad.sh:1406: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest) -./mad.sh:1407: if [ "${DebugAllEnabled}" == "true" ] ; then -./mad.sh:1408: debugHtml "github" "lbf_inst_curlimp$j" "$response" -./mad.sh:1409: fi -./mad.sh:1410: if [ ! -z "$response" ]; then -./mad.sh:1411: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response") -./mad.sh:1412: latestBinaryDate=$(grep -oPi -m 1 '(?<== MaxDownloadRetries)) ; then -./mad.sh:1485: echo -e "${RED}| FAILED: Size mismatch after downloading${NC}" -./mad.sh:1486: exit 1 +./mad.sh:1483: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path" +./mad.sh:1484: received_file_size=0 +./mad.sh:1485: if [ -f "$file_path" ] ; then +./mad.sh:1486: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]') +./mad.sh:1487: fi +./mad.sh:1488: if ((received_file_size == file_size_bytes)) ; then +./mad.sh:1489: break +./mad.sh:1490: elif ((received_file_size < file_size_bytes)) ; then +./mad.sh:1491: if ((j >= MaxDownloadRetries)) ; then +./mad.sh:1492: echo -e "${RED}| FAILED: Size mismatch after downloading${NC}" +./mad.sh:1493: exit 1 -- -./mad.sh:1529: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest) -./mad.sh:1530: if [ "${DebugAllEnabled}" == "true" ] ; then -./mad.sh:1531: debugHtml "github" "lbf_inst_curlimp$j" "$response" -./mad.sh:1532: fi -./mad.sh:1533: if [ ! -z "$response" ]; then -./mad.sh:1534: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response") -./mad.sh:1535: latestBinaryDate=$(grep -oPi -m 1 '(?<== MaxDownloadRetries)) ; then -./mad.sh:1608: echo -e "${RED}| FAILED: Size mismatch after downloading${NC}" -./mad.sh:1609: exit 1 +./mad.sh:1606: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path" +./mad.sh:1607: received_file_size=0 +./mad.sh:1608: if [ -f "$file_path" ] ; then +./mad.sh:1609: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]') +./mad.sh:1610: fi +./mad.sh:1611: if ((received_file_size == file_size_bytes)) ; then +./mad.sh:1612: break +./mad.sh:1613: elif ((received_file_size < file_size_bytes)) ; then +./mad.sh:1614: if ((j >= MaxDownloadRetries)) ; then +./mad.sh:1615: echo -e "${RED}| FAILED: Size mismatch after downloading${NC}" +./mad.sh:1616: exit 1 -- -./mad.sh:1804: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') -./mad.sh:1805: echo -e "Files:" -./mad.sh:1806: echo -e "${BLUE}${fil}${NC}" -./mad.sh:1807: echo -e "" -./mad.sh:1808: echo -e "" -./mad.sh:1809: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" -./mad.sh:1810: echo -e "_________________________________________________________________________" -./mad.sh:1811: echo -e "$maud_http" -./mad.sh:1812: echo -e "" -./mad.sh:1813: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})" -./mad.sh:1814: echo -e "_________________________________________________________________________" +./mad.sh:1811: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') +./mad.sh:1812: echo -e "Files:" +./mad.sh:1813: echo -e "${BLUE}${fil}${NC}" +./mad.sh:1814: echo -e "" +./mad.sh:1815: echo -e "" +./mad.sh:1816: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" +./mad.sh:1817: echo -e "_________________________________________________________________________" +./mad.sh:1818: echo -e "$maud_http" +./mad.sh:1819: echo -e "" +./mad.sh:1820: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})" +./mad.sh:1821: echo -e "_________________________________________________________________________" -- -./mad.sh:1817: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" -./mad.sh:1818: echo -e "_________________________________________________________________________" -./mad.sh:1819: echo -e "$maud_torcurl" -./mad.sh:1820: echo -e "" -./mad.sh:1821: echo -e "" -./mad.sh:1822: done -./mad.sh:1823: else -./mad.sh:1824: cd "$ScriptDir" -./mad.sh:1825: readarray -d $'' arrFiles < <(find . -name "*.sh" -printf '%p\n' | sort -Vk1) -./mad.sh:1826: cd "$WorkDir" -./mad.sh:1827: readarray -d $'' arrFiles2 < <(find . -name "*.sh" -printf '%p\n' | sort -Vk1) +./mad.sh:1824: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" +./mad.sh:1825: echo -e "_________________________________________________________________________" +./mad.sh:1826: echo -e "$maud_torcurl" +./mad.sh:1827: echo -e "" +./mad.sh:1828: echo -e "" +./mad.sh:1829: done +./mad.sh:1830: else +./mad.sh:1831: cd "$ScriptDir" +./mad.sh:1832: readarray -d $'' arrFiles < <(find . -name "*.sh" -printf '%p\n' | sort -Vk1) +./mad.sh:1833: cd "$WorkDir" +./mad.sh:1834: readarray -d $'' arrFiles2 < <(find . -name "*.sh" -printf '%p\n' | sort -Vk1) -- -./mad.sh:1832: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') -./mad.sh:1833: echo -e "Files:" -./mad.sh:1834: echo -e "${BLUE}${fil}${NC}" -./mad.sh:1835: echo -e "" -./mad.sh:1836: echo -e "" -./mad.sh:1837: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" -./mad.sh:1838: echo -e "_________________________________________________________________________" -./mad.sh:1839: echo -e "$maud_http" -./mad.sh:1840: echo -e "" -./mad.sh:1841: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl \"${NC})" -./mad.sh:1842: echo -e "_________________________________________________________________________" +./mad.sh:1839: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') +./mad.sh:1840: echo -e "Files:" +./mad.sh:1841: echo -e "${BLUE}${fil}${NC}" +./mad.sh:1842: echo -e "" +./mad.sh:1843: echo -e "" +./mad.sh:1844: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" +./mad.sh:1845: echo -e "_________________________________________________________________________" +./mad.sh:1846: echo -e "$maud_http" +./mad.sh:1847: echo -e "" +./mad.sh:1848: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl \"${NC})" +./mad.sh:1849: echo -e "_________________________________________________________________________" -- -./mad.sh:1845: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" -./mad.sh:1846: echo -e "_________________________________________________________________________" -./mad.sh:1847: echo -e "$maud_torcurl" -./mad.sh:1848: echo -e "" -./mad.sh:1849: done -./mad.sh:1850: for fil in "${arrFiles2[@]}"; -./mad.sh:1851: do -./mad.sh:1852: maud_http=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei '(http|https):') -./mad.sh:1853: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl') -./mad.sh:1854: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') -./mad.sh:1855: echo -e "Files:" -./mad.sh:1856: echo -e "${BLUE}${fil}${NC}" -./mad.sh:1857: echo -e "" -./mad.sh:1858: echo -e "" -./mad.sh:1859: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" -./mad.sh:1860: echo -e "_________________________________________________________________________" -./mad.sh:1861: echo -e "$maud_http" -./mad.sh:1862: echo -e "" -./mad.sh:1863: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})" -./mad.sh:1864: echo -e "_________________________________________________________________________" +./mad.sh:1852: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" +./mad.sh:1853: echo -e "_________________________________________________________________________" +./mad.sh:1854: echo -e "$maud_torcurl" +./mad.sh:1855: echo -e "" +./mad.sh:1856: done +./mad.sh:1857: for fil in "${arrFiles2[@]}"; +./mad.sh:1858: do +./mad.sh:1859: maud_http=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei '(http|https):') +./mad.sh:1860: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl') +./mad.sh:1861: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl') +./mad.sh:1862: echo -e "Files:" +./mad.sh:1863: echo -e "${BLUE}${fil}${NC}" +./mad.sh:1864: echo -e "" +./mad.sh:1865: echo -e "" +./mad.sh:1866: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})" +./mad.sh:1867: echo -e "_________________________________________________________________________" +./mad.sh:1868: echo -e "$maud_http" +./mad.sh:1869: echo -e "" +./mad.sh:1870: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})" +./mad.sh:1871: echo -e "_________________________________________________________________________" -- -./mad.sh:1867: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" -./mad.sh:1868: echo -e "_________________________________________________________________________" -./mad.sh:1869: echo -e "$maud_torcurl" -./mad.sh:1870: echo -e "" -./mad.sh:1871: done -./mad.sh:1872: fi -./mad.sh:1873:} -./mad.sh:1874:madStatus() { -./mad.sh:1875: local InputFile="$1" -./mad.sh:1876: if [ "$arg1" == "status" ] ; then -./mad.sh:1877: clear +./mad.sh:1874: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})" +./mad.sh:1875: echo -e "_________________________________________________________________________" +./mad.sh:1876: echo -e "$maud_torcurl" +./mad.sh:1877: echo -e "" +./mad.sh:1878: done +./mad.sh:1879: fi +./mad.sh:1880:} +./mad.sh:1881:madStatus() { +./mad.sh:1882: local InputFile="$1" +./mad.sh:1883: if [ "$arg1" == "status" ] ; then +./mad.sh:1884: clear -- -./mad.sh:3192: file_header=$(tor_curl_request --insecure -m 8 -s -D - -o /dev/null \ -./mad.sh:3193: -H "Connection: keep-alive" \ -./mad.sh:3194: -w 'EffectiveUrl=%{url_effective}' \ -./mad.sh:3195: "$download_url") -./mad.sh:3196: else -./mad.sh:3197: printf "| Retrieving Head: attempt #$j" -./mad.sh:3198: rm -f "${WorkDir}/.temp/directhead" -./mad.sh:3199: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -L -s -i "$download_url" | -./mad.sh:3200: tee "${WorkDir}/.temp/directhead" & -./mad.sh:3201: sleep 6 -./mad.sh:3202: [ -s "${WorkDir}/.temp/directhead" ] -./mad.sh:3203: kill $! 2>/dev/null -./mad.sh:3204: ) -./mad.sh:3205: if [ ! -f "${WorkDir}/.flocks/${remote_url//[^a-zA-Z0-9]/}" ]; then -./mad.sh:3206: touch "${WorkDir}/.flocks/${remote_url//[^a-zA-Z0-9]/}" -./mad.sh:3207: fi -./mad.sh:3208: rm -f "${WorkDir}/.temp/directhead" -./mad.sh:3209: fi +./mad.sh:3199: file_header=$(tor_curl_request --insecure -m 8 -s -D - -o /dev/null \ +./mad.sh:3200: -H "Connection: keep-alive" \ +./mad.sh:3201: -w 'EffectiveUrl=%{url_effective}' \ +./mad.sh:3202: "$download_url") +./mad.sh:3203: else +./mad.sh:3204: printf "| Retrieving Head: attempt #$j" +./mad.sh:3205: rm -f "${WorkDir}/.temp/directhead" +./mad.sh:3206: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -L -s -i "$download_url" | +./mad.sh:3207: tee "${WorkDir}/.temp/directhead" & +./mad.sh:3208: sleep 6 +./mad.sh:3209: [ -s "${WorkDir}/.temp/directhead" ] +./mad.sh:3210: kill $! 2>/dev/null +./mad.sh:3211: ) +./mad.sh:3212: if [ ! -f "${WorkDir}/.flocks/${remote_url//[^a-zA-Z0-9]/}" ]; then +./mad.sh:3213: touch "${WorkDir}/.flocks/${remote_url//[^a-zA-Z0-9]/}" +./mad.sh:3214: fi +./mad.sh:3215: rm -f "${WorkDir}/.temp/directhead" +./mad.sh:3216: fi -- -./mad.sh:3327: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --output "$file_path" -./mad.sh:3328: rc=$? -./mad.sh:3329: if [ $rc -ne 0 ] ; then -./mad.sh:3330: printf "${RED}Download Failed (bad exit status).${NC}" -./mad.sh:3331: if [ -f ${file_path} ]; then -./mad.sh:3332: printf "${YELLOW} Partial removed...${NC}" -./mad.sh:3333: printf "\n\n" -./mad.sh:3334: rm -f "${file_path}" -./mad.sh:3335: else -./mad.sh:3336: printf "\n\n" -./mad.sh:3337: fi +./mad.sh:3334: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --output "$file_path" +./mad.sh:3335: rc=$? +./mad.sh:3336: if [ $rc -ne 0 ] ; then +./mad.sh:3337: printf "${RED}Download Failed (bad exit status).${NC}" +./mad.sh:3338: if [ -f ${file_path} ]; then +./mad.sh:3339: printf "${YELLOW} Partial removed...${NC}" +./mad.sh:3340: printf "\n\n" +./mad.sh:3341: rm -f "${file_path}" +./mad.sh:3342: else +./mad.sh:3343: printf "\n\n" +./mad.sh:3344: fi -- -./mad.sh:3371: tor_curl_request --insecure -L --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path" -./mad.sh:3372: else -./mad.sh:3373: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --continue-at - --output "$file_path" -./mad.sh:3374: fi -./mad.sh:3375: received_file_size=0 -./mad.sh:3376: if [ -f "$file_path" ] ; then -./mad.sh:3377: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]') -./mad.sh:3378: fi -./mad.sh:3379: if CheckNoHtml "$remote_url" "$filename" "$file_path" ; then -./mad.sh:3380: containsHtml=false -./mad.sh:3381: else -./mad.sh:3382: containsHtml=true -./mad.sh:3383: fi +./mad.sh:3378: tor_curl_request --insecure -L --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path" +./mad.sh:3379: else +./mad.sh:3380: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --continue-at - --output "$file_path" +./mad.sh:3381: fi +./mad.sh:3382: received_file_size=0 +./mad.sh:3383: if [ -f "$file_path" ] ; then +./mad.sh:3384: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]') +./mad.sh:3385: fi +./mad.sh:3386: if CheckNoHtml "$remote_url" "$filename" "$file_path" ; then +./mad.sh:3387: containsHtml=false +./mad.sh:3388: else +./mad.sh:3389: containsHtml=true +./mad.sh:3390: fi -- -./mad.sh:3571: response=$(tor_curl_upload --insecure -i \ -./mad.sh:3572: -H "Content-Type: multipart/form-data" \ -./mad.sh:3573: -F "key=" \ -./mad.sh:3574: -F "time=$jira_timeval" \ -./mad.sh:3575: -F "file=@${filepath}" \ -./mad.sh:3576: "${jira_PostUrlHost}") -./mad.sh:3577: else ./mad.sh:3578: response=$(tor_curl_upload --insecure -i \ ./mad.sh:3579: -H "Content-Type: multipart/form-data" \ ./mad.sh:3580: -F "key=" \ ./mad.sh:3581: -F "time=$jira_timeval" \ -./mad.sh:3582: -F "files[]=@${arrFiles[@]}" \ +./mad.sh:3582: -F "file=@${filepath}" \ ./mad.sh:3583: "${jira_PostUrlHost}") -./mad.sh:3584: fi -./mad.sh:3585: if [ "${DebugAllEnabled}" == "true" ] ; then -./mad.sh:3586: debugHtml "${filepath##*/}" "${_hostCode}_upload" "post_url: ${jira_PostUrlHost}"$'\n'"${response}" -./mad.sh:3587: fi -./mad.sh:3588: if grep -Eqi ' 200 ' <<< "${response}" ; then +./mad.sh:3584: else +./mad.sh:3585: response=$(tor_curl_upload --insecure -i \ +./mad.sh:3586: -H "Content-Type: multipart/form-data" \ +./mad.sh:3587: -F "key=" \ +./mad.sh:3588: -F "time=$jira_timeval" \ +./mad.sh:3589: -F "files[]=@${arrFiles[@]}" \ +./mad.sh:3590: "${jira_PostUrlHost}") +./mad.sh:3591: fi +./mad.sh:3592: if [ "${DebugAllEnabled}" == "true" ] ; then +./mad.sh:3593: debugHtml "${filepath##*/}" "${_hostCode}_upload" "post_url: ${jira_PostUrlHost}"$'\n'"${response}" +./mad.sh:3594: fi +./mad.sh:3595: if grep -Eqi ' 200 ' <<< "${response}" ; then diff --git a/hosts/anonfile.sh b/hosts/anonfile.sh old mode 100644 new mode 100755 index dc46f3c..7c2c9b0 --- a/hosts/anonfile.sh +++ b/hosts/anonfile.sh @@ -1,6 +1,6 @@ #! Name: anonfile.sh #! Author: kittykat -#! Version: 2024.12.26 +#! Version: 2024.12.28 #! Desc: Add support for downloading and processing of urls for a new host #! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder #! @@ -137,7 +137,7 @@ anon_FetchFileInfo() { fi if grep -Eqi 'input type="hidden" name="id" value="' <<< "$response"; then printf "\\n" - echo -e "${GREEN}| Captcha found.${NC}" + echo -e "${GREEN}| Post link found.${NC}" post_op=$(grep -oP '(?<=input type="hidden" name="op" value=").*(?=">)' <<< "$response") post_id=$(grep -oP '(?<=input type="hidden" name="id" value=").*(?=">)' <<< "$response") post_fname=$(grep -oP '(?<=input type="hidden" name="fname" value=").*(?=">)' <<< "$response") @@ -327,8 +327,8 @@ anon_FetchFileInfo() { fi fi done - echo -e "| Captcha countdown (10s)…" - sleep 10s + echo -e "| Captcha countdown (5s)…" + sleep 5s maxfetchretries=1 echo -e "${GREEN}# Fetching download url…${NC}" for ((i=1; i<=$maxfetchretries; i++)); do @@ -413,7 +413,12 @@ anon_FetchFileInfo() { fi if grep -Eqi '.*$)' <<< "$response") + download_url=$(urlencode_literal_grouped_case_urlendingonly "$download_url") + elif grep -Eqi '.*$)' <<< "$response") download_url=$(urlencode_literal_grouped_case_urlendingonly "$download_url") fi @@ -531,12 +536,15 @@ anon_FetchFileInfo() { #! ----------- (3) Fetch File / Download File Function --------------- # #! anon_GetFile() { - echo -e "${GREEN}# Downloading…" + echo -e "${GREEN}# Downloading… ${BLUE}(No Resume)${NC}" echo -e "${YELLOW}| File path:${NC}\t./.inflight/${filename}\n" fileCnt=$1 retryCnt=$2 finalAttempt=$3 flockDownload="${WorkDir}/.flocks/${filename//[^a-zA-Z0-9\.\_\-]/}.flock" + if [ -f "$file_path" ]; then + rm -f "$file_path" + fi for ((j=1; j<=$MaxDownloadRetries; j++)); do pd_presize=0 if [ -f "$file_path" ] ; then @@ -552,12 +560,12 @@ anon_GetFile() { --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \ -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \ --referer "${fixed_url}" \ - "$download_url" --continue-at - --output "$file_path" + "$download_url" --output "$file_path" else tor_curl_request --insecure \ -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \ --referer "${fixed_url}" \ - "$download_url" --continue-at - --output "$file_path" + "$download_url" --output "$file_path" fi else if [ "${RateMonitorEnabled}" == "true" ]; then @@ -576,7 +584,7 @@ anon_GetFile() { -H "Sec-Fetch-User: ?1" \ -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \ --referer "${fixed_url}" \ - "$download_url" --continue-at - --output "$file_path" + "$download_url" --output "$file_path" else tor_curl_request --insecure \ -H "User-Agent: $RandomUA" \ @@ -592,7 +600,7 @@ anon_GetFile() { -H "Sec-Fetch-User: ?1" \ -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \ --referer "${fixed_url}" \ - "$download_url" --continue-at - --output "$file_path" + "$download_url" --output "$file_path" fi fi received_file_size=0 diff --git a/hosts/dashfile.sh b/hosts/dashfile.sh old mode 100644 new mode 100755 index a4c549e..f37093e --- a/hosts/dashfile.sh +++ b/hosts/dashfile.sh @@ -1,6 +1,6 @@ #! Name: dashfile.sh #! Author: kittykat -#! Version: 2024.12.25 +#! Version: 2025.01.02 #! Desc: Add support for downloading and processing of urls for a new host #! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder #! @@ -107,7 +107,7 @@ dash_FetchFileInfo() { continue fi fi - if grep -Eqi 'File Not Found|No such file with this filename|File was deleted|' <<< "$response"; then printf "\\n" echo -e "${GREEN}| Post link found.${NC}" @@ -194,7 +202,15 @@ dash_FetchFileInfo() { exitDownloadError=true removedDownload "${remote_url}" return 1 - fi + fi + if grep -Eqi 'This file reached max downloads limit' <<< "$response"; then + rm -f "${dash_cookie_jar}"; + printf "\\n" + echo -e "${RED}| The file has reached max downloads limit${NC}" + exitDownloadError=true + removedDownload "${remote_url}" + return 1 + fi if grep -Eqi 'you have to wait|seconds till next download' <<< "$response"; then if [ $i == $maxfetchretries ] ; then rm -f "${dash_cookie_jar}"; @@ -220,9 +236,21 @@ dash_FetchFileInfo() { pval4=$(grep -oP -m 1 '&#\K.*?(?=;)' <<< "$codeline" ) val1=$((pval1-48)); val2=$((pval2-48)); val3=$((pval3-48)); val4=$((pval4-48)) captcha_code="${val1}${val2}${val3}${val4}" - if grep -Eqi 'name="method_free" value="Free Download">' <<< "$response"; then + if [ -z "$captcha_code" ] || grep -Eqi '-' <<< "$captcha_code"; then + rm -f "${dash_cookie_jar}"; + printf "\\n" + echo -e "${RED}| Bad captcha code [2]${NC}" + warnAndRetryUnknownError=true + if [ "${finalAttempt}" == "true" ] ; then + failedRetryDownload "${remote_url}" "Bad captcha code [2]" "" + fi + return 1 + else printf "\\n" echo -e "${GREEN}| Captcha found.${NC}" + fi + if grep -Eqi 'name="method_free" value="Free Download">' <<< "$response"; then + echo -e "${GREEN}| Post2 found.${NC}" post_op=$(grep -oP '(?<=input type="hidden" name="op" value=").*(?=">)' <<< "$response") post_id=$(grep -oP '(?<=input type="hidden" name="id" value=").*(?=">)' <<< "$response") post_rand=$(grep -oP '(?<=input type="hidden" name="rand" value=").*(?=">)' <<< "$response") @@ -231,7 +259,6 @@ dash_FetchFileInfo() { else if [ $i == $maxfetchretries ] ; then rm -f "${dash_cookie_jar}"; - printf "\\n" echo -e "${RED}| Failed to extract download link [5]${NC}" warnAndRetryUnknownError=true if [ "${finalAttempt}" == "true" ] ; then @@ -243,25 +270,9 @@ dash_FetchFileInfo() { continue fi fi - if grep -Eqi '-' <<< "$captcha_code"; then - if [ $i == $maxfetchretries ] ; then - rm -f "${dash_cookie_jar}"; - printf "\\n" - echo -e "${RED}| Bad captcha code [2]${NC}" - warnAndRetryUnknownError=true - if [ "${finalAttempt}" == "true" ] ; then - failedRetryDownload "${remote_url}" "Bad captcha code [2]" "" - fi - return 1 - else - tor_identity="${RANDOM}" - continue - fi - fi if [[ -z "$captcha_code" ]] || [[ -z "$post_op" ]] || [[ -z "$post_id" ]] || [[ -z "$post_rand" ]] ; then if [ $i == $maxfetchretries ] ; then rm -f "${dash_cookie_jar}"; - printf "\\n" echo -e "${RED}| Failed to extract download link [6]${NC}" warnAndRetryUnknownError=true if [ "${finalAttempt}" == "true" ] ; then diff --git a/hosts/eternalhosting.sh b/hosts/eternalhosting.sh old mode 100644 new mode 100755 index 7b395b2..903ca30 --- a/hosts/eternalhosting.sh +++ b/hosts/eternalhosting.sh @@ -1,6 +1,6 @@ #! Name: eternalhosting.sh #! Author: kittykat -#! Version: 2024.09.13 +#! Version: 2024.12.28 #! Desc: Add support for downloading and processing of urls for a new host #! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder #! @@ -15,7 +15,7 @@ #! HostDomainRegex: The regex used to verify matching urls HostCode='eh' HostNick='eternal hosting' -HostFuncPrefix='direct' +HostFuncPrefix='eh' HostUrls='eternalcbrzpicytj4zyguygpmkjlkddxob7tptlr25cdipe5svyqoqd.onion' HostDomainRegex='^(http|https)://eternalcbrzpicytj4zyguygpmkjlkddxob7tptlr25cdipe5svyqoqd\.onion/file/' #! @@ -26,6 +26,21 @@ ListHostAndDomainRegexes=${ListHostAndDomainRegexes}'/'${HostCode}'/'${HostNick} #! #! ------------ (1) Host Main Download Function --------------- # #! -#! This is a direct= download host, so all the functions are already in mad.sh -#! Since the HostFuncPrefix is defined above as "direct", nothing further needs to be done as it will -#! call the direct_DownloadFile() function already in mad.sh +#! @REQUIRED: Host Main Download function +#! Must be named specifically as such: +#! _DownloadFile() +eh_DownloadFile() { + local pUrl="$1" + local pFileCnt="$2" + local pUrlMod="$pUrl" + if grep -Eqi '\.onion' <<< "$pUrlMod" && grep -Eqi 'https://' <<< "$pUrlMod" ; then + echo -e "${PINK}| Reverting .onion address to http...${NC}" + pUrlMod="${pUrlMod/https/http}" + fi + if grep -Eqi '\.onion/file/' <<< "$pUrlMod" ; then + echo -e "${PINK}| Switching to download url...${NC}" + pUrlMod="${pUrlMod/\.onion\/file\//\.onion\/download\/}" + fi + echo -e "[${BLUE}ModifiedUrl${NC}]: ${pUrlMod}" + direct_DownloadFile "$pUrl" "$pFileCnt" "$pUrlMod" +} diff --git a/hosts/fileblade.sh b/hosts/fileblade.sh old mode 100644 new mode 100755 index 4ddae54..87dd105 --- a/hosts/fileblade.sh +++ b/hosts/fileblade.sh @@ -1,6 +1,6 @@ #! Name: isupload.sh #! Author: kittykat -#! Version: 2024.12.20 +#! Version: 2024.12.28 #! Desc: Add support for downloading and processing of urls for a new host #! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder #! @@ -136,7 +136,6 @@ fb_FetchFileInfo() { post_id=$(grep -oPi -m 1 '(?<=input type="hidden" name="id" value=").*(?=">.*$)' <<< "$response") post_fname=$(grep -oPi -m 1 '(?<=input type="hidden" name="fname" value=").*(?=">.*$)' <<< "$response") post_referer=$(grep -oPi -m 1 '(?<=input type="hidden" name="referer" value=").*(?=">.*$)' <<< "$response") - post_action=$(urlencode_literal_grouped_case_urlendingonly "${post_action}") fi if [[ -z "$post_action" ]] || [[ -z "$post_op" ]] || [[ -z "$post_id" ]] || [[ -z "$post_fname" ]] ; then rm -f "${fb_cookie_jar}"; @@ -155,20 +154,18 @@ fb_FetchFileInfo() { break fi done - echo -e "| Download countdown (10s)…" - sleep 10s echo -e "${GREEN}# Fetching download2…${NC}" for ((i=1; i<=$maxfetchretries; i++)); do printf " _" download_url="" CLEANSTRING=${remote_url//[^a-zA-Z0-9]/} trap "rm -f ${WorkDir}/.flocks/${CLEANSTRING}; echo ""; rm -f $fb_cookie_jar; tput cnorm; exit" 0 1 2 3 6 15 - form_data="op=$post_op&usr_login=$post_usr_login&id=$post_id&fname=$post_fname&referer=$post_referer&method_free=method_free" + form_data="op=$post_op&usr_login=&id=$post_id&fname=$post_fname&referer=&method_free=method_free" response=$(tor_curl_request --insecure -L -s -X POST \ -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \ --data "$form_data" "$post_action") if [ "${DebugAllEnabled}" == "true" ] ; then - debugHtml "${remote_url##*/}" "fb_post" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}" + debugHtml "${remote_url##*/}" "fb_post(1)" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}" fi if [[ -z $response ]] ; then if [ $i == $maxfetchretries ] ; then @@ -201,14 +198,29 @@ fb_FetchFileInfo() { failedRetryDownload "${remote_url}" "Pro download only. [Free users not allowed download over 100MB]" "" return 1 fi - if grep -Eqi 'Just a moment...' <<< "$response"; then + if grep -Eqi 'Your subsequent download will be started in' <<< "$response"; then if [ $i == $maxfetchretries ] ; then rm -f "${fb_cookie_jar}"; printf "\\n" - echo -e "${RED}| Failed to extract download link [3].${NC}" + echo -e "${RED}| Subsequent download wait.. [3b]${NC}" warnAndRetryUnknownError=true if [ "${finalAttempt}" == "true" ] ; then - failedRetryDownload "${remote_url}" "" "" + failedRetryDownload "${remote_url}" "Subsequent download wait.. [3b]" "" + fi + return 1 + else + tor_identity="${RANDOM}" + continue + fi + fi + if grep -Eqi '

' <<< "$response"; then + if [ $i == $maxfetchretries ] ; then + rm -f "${fb_cookie_jar}"; + printf "\\n" + echo -e "${RED}| Failed to extract download link (Unknown warning encountered) [3c]${NC}" + warnAndRetryUnknownError=true + if [ "${finalAttempt}" == "true" ] ; then + failedRetryDownload "${remote_url}" "Unknown warning encountered in download2 [3c]" "" fi return 1 else @@ -255,7 +267,7 @@ fb_FetchFileInfo() { -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \ --data "$form_data" "$post_action") if [ "${DebugAllEnabled}" == "true" ] ; then - debugHtml "${remote_url##*/}" "fb_post2" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}" + debugHtml "${remote_url##*/}" "fb_post(2)" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}" fi if [[ -z $response ]] ; then if [ $i == $maxfetchretries ] ; then diff --git a/hosts/gofile.sh b/hosts/gofile.sh old mode 100644 new mode 100755 index e4c8de8..0326ccd --- a/hosts/gofile.sh +++ b/hosts/gofile.sh @@ -1,6 +1,6 @@ #! Name: gofile.sh #! Author: kittykat -#! Version: 2024.09.13 +#! Version: 2025.01.03 #! Desc: Add support for downloading and processing of urls for a new host #! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder #! @@ -197,6 +197,14 @@ gofile_FetchFileInfo() { else continue fi + fi + if grep -Eqi '"children":\{\}' <<< "$response"; then + rm -f "${gofile_cookie_jar}"; + printf "\\n" + echo -e "${RED}| Bulk download is a Premium feature. (No children)${NC}" + exitDownloadError=true + removedDownload "${remote_url}" "Bulk download is a Premium feature" + return 1 fi if grep -Eqi '"status":"ok"' <<< "$response"; then download_url=$(grep -oPi '(?<="link":").*?(?=")' <<< "$response") @@ -225,7 +233,7 @@ gofile_FetchFileInfo() { return 1 else continue - fi + fi fi else rm -f "${gofile_cookie_jar}"; diff --git a/hosts/up_axfc.sh b/hosts/up_axfc.sh old mode 100644 new mode 100755 index 11a2489..0419285 --- a/hosts/up_axfc.sh +++ b/hosts/up_axfc.sh @@ -1,6 +1,6 @@ #! Name: up_axfc.sh #! Author: kittykat -#! Version: 2024.12.26 +#! Version: 2025.01.02 #! Desc: Add support for uploading files to a new host #! Info: Files are accessible at https://www.axfc.net/ #! MaxSize: 2GB @@ -190,14 +190,28 @@ axfc_PostFile() { if [ "${DebugAllEnabled}" == "true" ] ; then debugHtml "${filepath##*/}" "${_hostCode}_upload" "post_url: ${PostUrlHost}"$'\n'"${response}" fi - if grep -Eqi 'Axfc Uploader -投稿完了.*キーワード付きURL:.*a href="https://www.axfc.net.*(QueryString無しVer)' <<< "${response}" ; then + response_ascii=$(echo "$response" | iconv -c -f UTF-8 -t ASCII//TRANSLIT) + if [ "${DebugAllEnabled}" == "true" ] ; then + debugHtml "${filepath##*/}" "${_hostCode}_upload_ascii" "post_url: ${PostUrlHost}"$'\n'"${response_ascii}" + fi + if grep -Eqi -m 1 'a href="https://www\.axfc\.net\/u\/.*\?key=1234"' <<< "${response_ascii}" ; then + url=$(grep -oPi -m 1 '(?<=a href="https://www\.axfc\.net\/u\/).*?(?=".*$)' <<< "$response_ascii") + key=$(grep -oPi -m 1 '(?<=\?key=).*?(?=".*$)' <<< "$response_ascii") + filesize=$(GetFileSize "$filepath" "false") + downloadLink="https://www.axfc.net/u/${url%%$'\n'*}?key=${key%%$'\n'*}" + echo -e "${GREEN}| Upload Success${NC}" + echo -e "| Size: ${BLUE}${filesize}${NC} bytes${NC}" + echo -e "| Link [1]: ${YELLOW}${downloadLink}${NC}" + successUpload "$pline" "${filepath}" "${_hostCode}" "${filesize}" "${downloadLink}" "{$response}" + return 0 + elif grep -Eqi 'Axfc Uploader -投稿完了.*キーワード付きURL:.*a href="https://www.axfc.net.*(QueryString無しVer)' <<< "${response}" ; then subSearch=$(awk '/Axfc Uploader -投稿完了/,/(QueryString無しVer)/' <<< "$response") url=$(grep -oPi -m 1 '(?<=キーワード付きURL:\\\\\ファイルIDのみ.*$)' <<< "$subSearch") filesize=$(GetFileSize "$filepath" "false") downloadLink="${url%%$'\n'*}" echo -e "${GREEN}| Upload Success${NC}" echo -e "| Size: ${BLUE}${filesize}${NC} bytes${NC}" - echo -e "| Link: ${YELLOW}${downloadLink}${NC}" + echo -e "| Link [2]: ${YELLOW}${downloadLink}${NC}" successUpload "$pline" "${filepath}" "${_hostCode}" "${filesize}" "${downloadLink}" "{$response}" return 0 else diff --git a/mad.sh b/mad.sh index e77396d..6cf2f4d 100644 --- a/mad.sh +++ b/mad.sh @@ -1,6 +1,6 @@ #!/bin/bash -# Copyright 2023-2024 kittykat (kittykat@morke.org) | (luvherpurr@torbox3uiot6wchz.onion) +# Copyright 2023-2025 kittykat (kittykat@morke.org) | (luvherpurr@torbox3uiot6wchz.onion) # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. @@ -31,9 +31,16 @@ # * klonkerz - feedback and suggestions, url only processing # * Everyone who provided feedback and helped test.. and those who wish to remain anonymous -ScriptVersion=2024.12.26 +ScriptVersion=2025.01.04 #================================================= # Recent Additions +# 2024.01.03 - [gofile] Detect "Bulk download is a Premium feature" response (no children) +# 2025.01.02 - [up_axfc] Update PUT response check to handle kanji chars (remove) +# 2025.01.02 - [dashfile] Add response 'This file reached max downloads limit'. New cookie on captcha fail +# 2024.12.28 - [dashfile] Update captcha code check +# 2024.12.28 - [anonfile] Add new download link href response +# 2024.12.28 - [fileblade] Add additional response handling (subsequent downloads, unknown warnings) +# 2024.12.28 - [eternalhosting] Update eternal.onion to handle landing page (eternal.onion/file/) # 2024.12.26 - [up_kouploader / up_axfc / up_torup] Fixed failedRetryUpload (was using download logging) # * Thanks Belky # 2024.12.26 - [anonfile / up_anonfile] Add anonfile.de as upload / download host