# 2024.01.03 - [gofile] Detect "Bulk download is a Premium feature" response (no children)

# 2025.01.02 - [up_axfc] Update PUT response check to handle kanji chars (remove)
# 2025.01.02 - [dashfile] Add response 'This file reached max downloads limit'. New cookie on captcha fail
# 2024.12.28 - [dashfile] Update captcha code check
# 2024.12.28 - [anonfile] Add new download link href response
# 2024.12.28 - [fileblade] Add additional response handling (subsequent downloads, unknown warnings)
# 2024.12.28 - [eternalhosting] Update eternal.onion to handle landing page (eternal.onion/file/)
This commit is contained in:
kittykat 2025-01-04 05:51:50 +00:00
parent e6804e01e1
commit 30eedaf567
Signed by: kittykat
GPG key ID: E3F1556620F70C3C
10 changed files with 766 additions and 687 deletions

View file

@ -1,4 +1,4 @@
DateTime: 24.12.26
DateTime: 25.01.04
Files:
./hosts/1fichier.sh
@ -163,12 +163,12 @@ _________________________________________________________________________
./hosts/anonfile.sh:186: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/anonfile.sh:240: tor_curl_request --insecure -s "$captcha_img_url" --output "$tmp_captcha_img"
./hosts/anonfile.sh:340: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/anonfile.sh:446: file_header=$(tor_curl_request -i -s --head \
./hosts/anonfile.sh:549: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./hosts/anonfile.sh:551: tor_curl_request --insecure \
./hosts/anonfile.sh:557: tor_curl_request --insecure \
./hosts/anonfile.sh:564: tor_curl_request --insecure \
./hosts/anonfile.sh:581: tor_curl_request --insecure \
./hosts/anonfile.sh:451: file_header=$(tor_curl_request -i -s --head \
./hosts/anonfile.sh:557: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./hosts/anonfile.sh:559: tor_curl_request --insecure \
./hosts/anonfile.sh:565: tor_curl_request --insecure \
./hosts/anonfile.sh:572: tor_curl_request --insecure \
./hosts/anonfile.sh:589: tor_curl_request --insecure \
./hosts/anonsharing.sh:91: response=$(tor_curl_request --insecure -i -s \
./hosts/anonsharing.sh:150: file_header=$(tor_curl_request --insecure --head -L -i -s \
./hosts/anonsharing.sh:158: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -s -i \
@ -219,14 +219,14 @@ _________________________________________________________________________
./hosts/dailyuploads.sh:511: tor_curl_request --insecure \
./hosts/dailyuploads.sh:528: tor_curl_request --insecure \
./hosts/dashfile.sh:90: response=$(tor_curl_request --insecure -L -s -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:169: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/dashfile.sh:297: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/dashfile.sh:386: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url")
./hosts/dashfile.sh:484: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./hosts/dashfile.sh:486: tor_curl_request --insecure \
./hosts/dashfile.sh:491: tor_curl_request --insecure \
./hosts/dashfile.sh:177: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/dashfile.sh:308: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/dashfile.sh:397: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url")
./hosts/dashfile.sh:495: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./hosts/dashfile.sh:497: tor_curl_request --insecure \
./hosts/dashfile.sh:513: tor_curl_request --insecure \
./hosts/dashfile.sh:502: tor_curl_request --insecure \
./hosts/dashfile.sh:508: tor_curl_request --insecure \
./hosts/dashfile.sh:524: tor_curl_request --insecure \
./hosts/dataupload.sh:90: response=$(tor_curl_request --insecure -L -s -b "${dup_cookie_jar}" -c "${dup_cookie_jar}" "$remote_url")
./hosts/dataupload.sh:166: response=$(tor_curl_request --insecure -svo. -X POST \
./hosts/dataupload.sh:234: file_header=$(tor_curl_request --insecure -L --head -s "$download_url")
@ -255,14 +255,14 @@ _________________________________________________________________________
./hosts/examples/ExampleNewHost.sh:201: tor_curl_request --insecure --referer "$download_url" "$download_url" --continue-at - --output "$file_path"
./hosts/examples/up_example.sh:112: response=$(tor_curl_upload --insecure \
./hosts/fileblade.sh:90: response=$(tor_curl_request --insecure -L -s -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" "$remote_url")
./hosts/fileblade.sh:167: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/fileblade.sh:254: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/fileblade.sh:323: file_header=$(tor_curl_request --insecure -L --head -s "$download_url")
./hosts/fileblade.sh:438: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./hosts/fileblade.sh:440: tor_curl_request --insecure -L \
./hosts/fileblade.sh:444: tor_curl_request --insecure -L \
./hosts/fileblade.sh:449: tor_curl_request --insecure \
./hosts/fileblade.sh:464: tor_curl_request --insecure \
./hosts/fileblade.sh:164: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/fileblade.sh:266: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/fileblade.sh:335: file_header=$(tor_curl_request --insecure -L --head -s "$download_url")
./hosts/fileblade.sh:450: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./hosts/fileblade.sh:452: tor_curl_request --insecure -L \
./hosts/fileblade.sh:456: tor_curl_request --insecure -L \
./hosts/fileblade.sh:461: tor_curl_request --insecure \
./hosts/fileblade.sh:476: tor_curl_request --insecure \
./hosts/fileditch.sh:85: file_header=$(tor_curl_request --insecure --head -L -s "$download_url")
./hosts/fileditch.sh:176: tor_curl_request --insecure --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path"
./hosts/fileditch.sh:178: tor_curl_request --insecure "$download_url" --continue-at - --output "$file_path"
@ -282,9 +282,9 @@ _________________________________________________________________________
./hosts/firestorage.sh:337: tor_curl_request --insecure -L -G "$download_url" --continue-at - --output "$file_path"
./hosts/gofile.sh:97: response=$(tor_curl_request --insecure -s -X POST \
./hosts/gofile.sh:170: response=$(tor_curl_request --insecure -G -L -s \
./hosts/gofile.sh:250: file_header=$(tor_curl_request --insecure -L --head -s \
./hosts/gofile.sh:369: tor_curl_request --insecure -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/gofile.sh:383: tor_curl_request --insecure -G \
./hosts/gofile.sh:258: file_header=$(tor_curl_request --insecure -L --head -s \
./hosts/gofile.sh:377: tor_curl_request --insecure -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/gofile.sh:391: tor_curl_request --insecure -G \
./hosts/hexload.sh:108: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download")
./hosts/hexload.sh:116: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download")
./hosts/hexload.sh:122: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download")
@ -452,119 +452,119 @@ _________________________________________________________________________
./hosts/youdbox.sh:183: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url")
./hosts/youdbox.sh:276: tor_curl_request --insecure -L -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path"
./hosts/youdbox.sh:278: tor_curl_request --insecure -L -G "$download_url" --continue-at - --output "$file_path"
./mad.sh:100:UseTorCurlImpersonate=false
./mad.sh:391:tor_curl_request() {
./mad.sh:392: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:393: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:395: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:398:tor_curl_request_extended() {
./mad.sh:400: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:401: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@"
./mad.sh:403: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@"
./mad.sh:406:tor_curl_upload() {
./mad.sh:107:UseTorCurlImpersonate=false
./mad.sh:398:tor_curl_request() {
./mad.sh:399: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:400: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:402: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:405:tor_curl_request_extended() {
./mad.sh:407: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:409: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval --compressed --globoff "$@"
./mad.sh:411: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:415: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@"
./mad.sh:417: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@"
./mad.sh:1361:install_curl_impersonate() {
./mad.sh:1363: echo -e "${BLUE}lwthiker curl_impersonate${NC} is the original dev, but it is relatively inactive."
./mad.sh:1364: echo -e "- Currently uses curl v8.1.1."
./mad.sh:1368: echo -e "${BLUE}lexiforest curl_impersonate${NC} is an active fork of curl_impersonate."
./mad.sh:1369: echo -e "+ Currently uses curl v8.7.1"
./mad.sh:1373: PS3='Please select which curl_impersonate to install: '
./mad.sh:1381: install_curl_impersonate_lwthiker_orig
./mad.sh:1385: install_curl_impersonate_lexiforest_fork
./mad.sh:1395:install_curl_impersonate_lwthiker_orig() {
./mad.sh:1399: echo -e "${BLUE}lwthiker curl_impersonate${NC} is the original curl_impersonate."
./mad.sh:1400: echo -e "+ Currently uses curl v8.1.1, and has low activity for updates"
./mad.sh:1403: echo -e "${GREEN}| Fetching:${NC} latest ${BLUE}lwthiker curl_impersonate${NC} info from github...${NC}"
./mad.sh:1406: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest)
./mad.sh:1408: debugHtml "github" "lbf_inst_curlimp$j" "$response"
./mad.sh:1411: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response")
./mad.sh:1421: yes_or_no "Do you wish to download and extract curl_impersonate $latestTag (using tor+curl)?" && {
./mad.sh:1423: download_url='https://github.com/lwthiker/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz'
./mad.sh:1426: file_header=$(tor_curl_request --insecure --head -Ls "$download_url")
./mad.sh:1428: debugHtml "github" "head_inst_curlimp$j" "${file_header}"
./mad.sh:1476: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path"
./mad.sh:1505: echo -e "| Extracting curl_impersonate..."
./mad.sh:1507: rm -f "${ScriptDir}"/curl*
./mad.sh:1508: mv "$extract_location/curl-impersonate-ff" "${ScriptDir}/"
./mad.sh:1509: mv "$extract_location/curl_ff109" "${ScriptDir}/"
./mad.sh:1510: echo -e "${GREEN}| Done.${NC} Update ${BLUE}\"UseTorCurlImpersonate=true\"${NC} in script to use..."
./mad.sh:1518:install_curl_impersonate_lexiforest_fork() {
./mad.sh:1522: echo -e "${BLUE}lexiforest curl_impersonate${NC} is an active fork of curl_impersonate."
./mad.sh:1523: echo -e "+ Currently uses curl v8.7.1, and is patched for latest CVEs"
./mad.sh:1526: echo -e "${GREEN}| Fetching:${NC} latest ${BLUE}lexiforest curl_impersonate fork${NC} info from github...${NC}"
./mad.sh:1529: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest)
./mad.sh:1531: debugHtml "github" "lbf_inst_curlimp$j" "$response"
./mad.sh:1534: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response")
./mad.sh:1544: yes_or_no "Do you wish to download and extract curl_impersonate $latestTag (using tor+curl)?" && {
./mad.sh:1546: download_url='https://github.com/lexiforest/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz'
./mad.sh:1549: file_header=$(tor_curl_request --insecure --head -Ls "$download_url")
./mad.sh:1551: debugHtml "github" "head_inst_curlimp$j" "${file_header}"
./mad.sh:1599: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path"
./mad.sh:1628: echo -e "| Extracting curl_impersonate..."
./mad.sh:1630: rm -f "${ScriptDir}"/curl*
./mad.sh:1631: mv "$extract_location/curl-impersonate-chrome" "${ScriptDir}/"
./mad.sh:1632: mv "$extract_location/curl_chrome131" "${ScriptDir}/"
./mad.sh:1633: echo -e "${GREEN}| Done.${NC} Update ${BLUE}\"UseTorCurlImpersonate=true\"${NC} in script to use..."
./mad.sh:1795: echo -e ":${NC} ${GREEN}MAD${PINK} Audit${NC} : Reports usage of http & curl in scripts${PINK}${BLD} :"
./mad.sh:1803: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl')
./mad.sh:1804: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1813: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})"
./mad.sh:1815: echo -e "$maud_curl"
./mad.sh:1817: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1819: echo -e "$maud_torcurl"
./mad.sh:1831: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl')
./mad.sh:1832: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1841: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl \"${NC})"
./mad.sh:1843: echo -e "$maud_curl"
./mad.sh:1845: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1847: echo -e "$maud_torcurl"
./mad.sh:1853: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl')
./mad.sh:1854: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1863: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})"
./mad.sh:1865: echo -e "$maud_curl"
./mad.sh:1867: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1869: echo -e "$maud_torcurl"
./mad.sh:2816: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:2817: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n"
./mad.sh:2819: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n"
./mad.sh:2991: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:2992: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n"
./mad.sh:2994: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n"
./mad.sh:3192: file_header=$(tor_curl_request --insecure -m 8 -s -D - -o /dev/null \
./mad.sh:3199: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -L -s -i "$download_url" |
./mad.sh:3327: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --output "$file_path"
./mad.sh:3371: tor_curl_request --insecure -L --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path"
./mad.sh:3373: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --continue-at - --output "$file_path"
./mad.sh:3571: response=$(tor_curl_upload --insecure -i \
./mad.sh:408: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@"
./mad.sh:410: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@"
./mad.sh:413:tor_curl_upload() {
./mad.sh:414: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:416: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval --compressed --globoff "$@"
./mad.sh:418: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:422: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@"
./mad.sh:424: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@"
./mad.sh:1368:install_curl_impersonate() {
./mad.sh:1370: echo -e "${BLUE}lwthiker curl_impersonate${NC} is the original dev, but it is relatively inactive."
./mad.sh:1371: echo -e "- Currently uses curl v8.1.1."
./mad.sh:1375: echo -e "${BLUE}lexiforest curl_impersonate${NC} is an active fork of curl_impersonate."
./mad.sh:1376: echo -e "+ Currently uses curl v8.7.1"
./mad.sh:1380: PS3='Please select which curl_impersonate to install: '
./mad.sh:1388: install_curl_impersonate_lwthiker_orig
./mad.sh:1392: install_curl_impersonate_lexiforest_fork
./mad.sh:1402:install_curl_impersonate_lwthiker_orig() {
./mad.sh:1406: echo -e "${BLUE}lwthiker curl_impersonate${NC} is the original curl_impersonate."
./mad.sh:1407: echo -e "+ Currently uses curl v8.1.1, and has low activity for updates"
./mad.sh:1410: echo -e "${GREEN}| Fetching:${NC} latest ${BLUE}lwthiker curl_impersonate${NC} info from github...${NC}"
./mad.sh:1413: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest)
./mad.sh:1415: debugHtml "github" "lbf_inst_curlimp$j" "$response"
./mad.sh:1418: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response")
./mad.sh:1428: yes_or_no "Do you wish to download and extract curl_impersonate $latestTag (using tor+curl)?" && {
./mad.sh:1430: download_url='https://github.com/lwthiker/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz'
./mad.sh:1433: file_header=$(tor_curl_request --insecure --head -Ls "$download_url")
./mad.sh:1435: debugHtml "github" "head_inst_curlimp$j" "${file_header}"
./mad.sh:1483: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path"
./mad.sh:1512: echo -e "| Extracting curl_impersonate..."
./mad.sh:1514: rm -f "${ScriptDir}"/curl*
./mad.sh:1515: mv "$extract_location/curl-impersonate-ff" "${ScriptDir}/"
./mad.sh:1516: mv "$extract_location/curl_ff109" "${ScriptDir}/"
./mad.sh:1517: echo -e "${GREEN}| Done.${NC} Update ${BLUE}\"UseTorCurlImpersonate=true\"${NC} in script to use..."
./mad.sh:1525:install_curl_impersonate_lexiforest_fork() {
./mad.sh:1529: echo -e "${BLUE}lexiforest curl_impersonate${NC} is an active fork of curl_impersonate."
./mad.sh:1530: echo -e "+ Currently uses curl v8.7.1, and is patched for latest CVEs"
./mad.sh:1533: echo -e "${GREEN}| Fetching:${NC} latest ${BLUE}lexiforest curl_impersonate fork${NC} info from github...${NC}"
./mad.sh:1536: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest)
./mad.sh:1538: debugHtml "github" "lbf_inst_curlimp$j" "$response"
./mad.sh:1541: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response")
./mad.sh:1551: yes_or_no "Do you wish to download and extract curl_impersonate $latestTag (using tor+curl)?" && {
./mad.sh:1553: download_url='https://github.com/lexiforest/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz'
./mad.sh:1556: file_header=$(tor_curl_request --insecure --head -Ls "$download_url")
./mad.sh:1558: debugHtml "github" "head_inst_curlimp$j" "${file_header}"
./mad.sh:1606: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path"
./mad.sh:1635: echo -e "| Extracting curl_impersonate..."
./mad.sh:1637: rm -f "${ScriptDir}"/curl*
./mad.sh:1638: mv "$extract_location/curl-impersonate-chrome" "${ScriptDir}/"
./mad.sh:1639: mv "$extract_location/curl_chrome131" "${ScriptDir}/"
./mad.sh:1640: echo -e "${GREEN}| Done.${NC} Update ${BLUE}\"UseTorCurlImpersonate=true\"${NC} in script to use..."
./mad.sh:1802: echo -e ":${NC} ${GREEN}MAD${PINK} Audit${NC} : Reports usage of http & curl in scripts${PINK}${BLD} :"
./mad.sh:1810: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl')
./mad.sh:1811: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1820: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})"
./mad.sh:1822: echo -e "$maud_curl"
./mad.sh:1824: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1826: echo -e "$maud_torcurl"
./mad.sh:1838: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl')
./mad.sh:1839: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1848: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl \"${NC})"
./mad.sh:1850: echo -e "$maud_curl"
./mad.sh:1852: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1854: echo -e "$maud_torcurl"
./mad.sh:1860: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl')
./mad.sh:1861: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1870: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})"
./mad.sh:1872: echo -e "$maud_curl"
./mad.sh:1874: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1876: echo -e "$maud_torcurl"
./mad.sh:2823: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:2824: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n"
./mad.sh:2826: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n"
./mad.sh:2998: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:2999: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n"
./mad.sh:3001: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n"
./mad.sh:3199: file_header=$(tor_curl_request --insecure -m 8 -s -D - -o /dev/null \
./mad.sh:3206: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -L -s -i "$download_url" |
./mad.sh:3334: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --output "$file_path"
./mad.sh:3378: tor_curl_request --insecure -L --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path"
./mad.sh:3380: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --continue-at - --output "$file_path"
./mad.sh:3578: response=$(tor_curl_upload --insecure -i \
./mad.sh:3649:if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:3650: curl_impersonate=()
./mad.sh:3651: readarray -d $'' arrFiles < <(find "$ScriptDir" -maxdepth 1 -name "curl_*" -printf '%p\n' | sort -Vk1)
./mad.sh:3652: bFoundCurlHeader=false
./mad.sh:3656: curl_impersonate=($fil)
./mad.sh:3657: bFoundCurlHeader=true
./mad.sh:3661: if [ "$bFoundCurlHeader" == "false" ]; then
./mad.sh:3662: echo -e "${RED}[ERROR] Missing dependency \"curl-impersonate\"!${NC}"
./mad.sh:3665: echo -e "You'll need to download ${GREEN}\"curl-impersonate\"${NC}."
./mad.sh:3668: echo -e "The latest binary can be obtained on GitHub, search for \"curl-impersonate\""
./mad.sh:3670: echo -e " 1. Visit the page of curl-impersonate and add \"/releases/latest/\" at end of URL."
./mad.sh:3674: echo -e " 4. Download archive ${GREEN}\"curl-impersonate-vX.Y.Z.x86_64-linux-gnu.tar.gz\"${YELLOW}."
./mad.sh:3675: echo -e " 5. Extract files ${GREEN}\"curl-impersonate-ff\"${NC} and ${GREEN}\"curl_ff109\"${NC} next to this script."
./mad.sh:3678: echo -e "run $0 install_curl_impersonate\\n"
./mad.sh:3680: yes_or_no "Do you wish to download and extract latest curl_impersonate (using tor+curl)?" && {
./mad.sh:3681: UseTorCurlImpersonate=false
./mad.sh:3682: install_curl_impersonate
./mad.sh:3766: echo -e "[${YELLOW}Install curl_impersonate${NC}]: Downloads the latest binary for curl_impersonate from github repo (3 choices)"
./mad.sh:3767: printf " %s install_curl_impersonate\\n" "$0"
./mad.sh:3845:elif [[ "$arg1" == "install_curl_impersonate" ]]; then
./mad.sh:3846: install_curl_impersonate
./mad.sh:3877:if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:3878: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n"
./mad.sh:3880: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n"
./mad.sh:3585: response=$(tor_curl_upload --insecure -i \
./mad.sh:3656:if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:3657: curl_impersonate=()
./mad.sh:3658: readarray -d $'' arrFiles < <(find "$ScriptDir" -maxdepth 1 -name "curl_*" -printf '%p\n' | sort -Vk1)
./mad.sh:3659: bFoundCurlHeader=false
./mad.sh:3663: curl_impersonate=($fil)
./mad.sh:3664: bFoundCurlHeader=true
./mad.sh:3668: if [ "$bFoundCurlHeader" == "false" ]; then
./mad.sh:3669: echo -e "${RED}[ERROR] Missing dependency \"curl-impersonate\"!${NC}"
./mad.sh:3672: echo -e "You'll need to download ${GREEN}\"curl-impersonate\"${NC}."
./mad.sh:3675: echo -e "The latest binary can be obtained on GitHub, search for \"curl-impersonate\""
./mad.sh:3677: echo -e " 1. Visit the page of curl-impersonate and add \"/releases/latest/\" at end of URL."
./mad.sh:3681: echo -e " 4. Download archive ${GREEN}\"curl-impersonate-vX.Y.Z.x86_64-linux-gnu.tar.gz\"${YELLOW}."
./mad.sh:3682: echo -e " 5. Extract files ${GREEN}\"curl-impersonate-ff\"${NC} and ${GREEN}\"curl_ff109\"${NC} next to this script."
./mad.sh:3685: echo -e "run $0 install_curl_impersonate\\n"
./mad.sh:3687: yes_or_no "Do you wish to download and extract latest curl_impersonate (using tor+curl)?" && {
./mad.sh:3688: UseTorCurlImpersonate=false
./mad.sh:3689: install_curl_impersonate
./mad.sh:3773: echo -e "[${YELLOW}Install curl_impersonate${NC}]: Downloads the latest binary for curl_impersonate from github repo (3 choices)"
./mad.sh:3774: printf " %s install_curl_impersonate\\n" "$0"
./mad.sh:3852:elif [[ "$arg1" == "install_curl_impersonate" ]]; then
./mad.sh:3853: install_curl_impersonate
./mad.sh:3884:if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:3885: printf "client: ${GREEN}Tor${NC} + ${BLUE}curl_impersonate${NC}\\n"
./mad.sh:3887: printf "client: ${GREEN}Tor${NC} + ${GREEN}curl${NC}\\n"
./plugins/pjscloud.sh:44: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./plugins/pjscloud.sh:45: response=$("${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" \
./plugins/pjscloud.sh:53: response=$(curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" \

View file

@ -1,4 +1,4 @@
DateTime: 24.12.26
DateTime: 25.01.04
Files:
./hosts/1fichier.sh
@ -150,7 +150,7 @@ _________________________________________________________________________
./hosts/9saves.sh:141: --data "$form_data" "https://9saves.com/")
./hosts/anonfile.sh:230: if grep -Eqi 'img src="https://anonfile.de/captchas/' <<< "$response" ; then
./hosts/anonfile.sh:414: if grep -Eqi '<a class="stretched-link" href="https://anonfile.de' <<< "$response"; then
./hosts/anonfile.sh:546: anon_host=$(grep -oPi '(?<=https://).*(?=/)' <<< "$fixed_url")
./hosts/anonfile.sh:554: anon_host=$(grep -oPi '(?<=https://).*(?=/)' <<< "$fixed_url")
./hosts/anonsharing.sh:95: "https://anonsharing.com/account/ajax/file_details")
./hosts/anonsharing.sh:121: if grep -Eqi 'openUrl\('"'"'https:\\/\\/anonsharing.com\\/' <<< "$response"; then
./hosts/anonsharing.sh:124: download_url='https://anonsharing.com/'$(grep -oPi '(?<=openUrl\('"'"'https:\\/\\/anonsharing.com\\/).*?(?='"'"'.*$)' <<< "$response" | head -1)
@ -169,8 +169,8 @@ _________________________________________________________________________
./hosts/dailyuploads.sh:125: if grep -Eqi 'img src="https://dailyuploads.net/captchas/' <<< "$response" ; then
./hosts/dailyuploads.sh:280: response=$(pjscloud_tor_request "https://hexload.com/download" "$form_data")
./hosts/dailyuploads.sh:493: daily_host=$(grep -oPi '(?<=https://).*(?=/)' <<< "$fixed_url")
./hosts/dashfile.sh:356: if grep -Eqi '<a href="https://(.*\.)?dashfile\.net/files/' <<< "$response"; then
./hosts/dashfile.sh:359: download_url=$(grep -oP '(?<=<a href=").*(?=">https://.*dashfile.net.*$)' <<< "$response")
./hosts/dashfile.sh:367: if grep -Eqi '<a href="https://(.*\.)?dashfile\.net/files/' <<< "$response"; then
./hosts/dashfile.sh:370: download_url=$(grep -oP '(?<=<a href=").*(?=">https://.*dashfile.net.*$)' <<< "$response")
./hosts/dataupload.sh:133: post_action="https://dataupload.net/"
./hosts/dataupload.sh:210: if ! grep -Eqi "location: https://dataupload.net/d/" <<< "$response"; then
./hosts/dosya.sh:168: if grep -Eqi 'https://dosyaupload.com' <<< "$dos_url" ; then
@ -179,10 +179,11 @@ _________________________________________________________________________
./hosts/downloadgg.sh:268: -H "Origin: https://download.gg" \
./hosts/downloadgg.sh:278: -H "Origin: https://download.gg" \
./hosts/downloadgg.sh:297: -H "Origin: https://download.gg" \
./hosts/eternalhosting.sh:36: if grep -Eqi '\.onion' <<< "$pUrlMod" && grep -Eqi 'https://' <<< "$pUrlMod" ; then
./hosts/examples/up_example.sh:105: local ar_HUP[0]='https://oshi.at'
./hosts/fileblade.sh:298: if ! grep -Eqi '<a href="https://de6.fileblade.com/files/' <<< "$response"; then
./hosts/fileblade.sh:310: download_url=$(grep -oP -m 1 '(?<=a href="https://de6.fileblade.com/files/).*?(?=" class=.*$)' <<< "$response")
./hosts/fileblade.sh:312: download_url='https://de6.fileblade.com/files/'$(urlencode_literal_grouped_case_urlendingonly "$download_url")
./hosts/fileblade.sh:310: if ! grep -Eqi '<a href="https://de6.fileblade.com/files/' <<< "$response"; then
./hosts/fileblade.sh:322: download_url=$(grep -oP -m 1 '(?<=a href="https://de6.fileblade.com/files/).*?(?=" class=.*$)' <<< "$response")
./hosts/fileblade.sh:324: download_url='https://de6.fileblade.com/files/'$(urlencode_literal_grouped_case_urlendingonly "$download_url")
./hosts/filedot.sh:119: "https://filedot.to/login.html")
./hosts/filedot.sh:160: -H "Origin: https://filedot.to" \
./hosts/filedot.sh:162: -H "Referer: https://filedot.to/login.html" \
@ -197,8 +198,8 @@ _________________________________________________________________________
./hosts/gofile.sh:108: "https://api.gofile.io/accounts")
./hosts/gofile.sh:183: "https://api.gofile.io/contents/$file_id")
./hosts/gofile.sh:185: debugHtml "${remote_url##*/}" "gofile_contents$i" "url: https://api.gofile.io/contents/${file_id}?${form_data}"$'\n'"${response}"
./hosts/gofile.sh:204: cnturls=$(grep -oin 'https://' <<< "$download_url" | wc -l)
./hosts/gofile.sh:303: cdn_url="https:"$(grep -oPi '(?<=location: ).*' <<< "$file_header")
./hosts/gofile.sh:212: cnturls=$(grep -oin 'https://' <<< "$download_url" | wc -l)
./hosts/gofile.sh:311: cdn_url="https:"$(grep -oPi '(?<=location: ).*' <<< "$file_header")
./hosts/hexload.sh:102: response=$(pjscloud_tor_request "https://hexload.com/download" "$form_data")
./hosts/hexload.sh:108: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download")
./hosts/hexload.sh:116: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download")
@ -247,7 +248,10 @@ _________________________________________________________________________
./hosts/up_axfc.sh:100: fixed_url='https://www.axfc.net/u/post_m.pl'
./hosts/up_axfc.sh:137: -H 'Referer: https://www.axfc.net/u/post_m.pl' \
./hosts/up_axfc.sh:139: -H 'Origin: https://www.axfc.net' \
./hosts/up_axfc.sh:193: if grep -Eqi 'Axfc Uploader -投稿完了.*キーワード付きURL.*a href="https://www.axfc.net.*(QueryString無しVer)' <<< "${response}" ; then
./hosts/up_axfc.sh:197: if grep -Eqi -m 1 'a href="https://www\.axfc\.net\/u\/.*\?key=1234"' <<< "${response_ascii}" ; then
./hosts/up_axfc.sh:198: url=$(grep -oPi -m 1 '(?<=a href="https://www\.axfc\.net\/u\/).*?(?=".*$)' <<< "$response_ascii")
./hosts/up_axfc.sh:201: downloadLink="https://www.axfc.net/u/${url%%$'\n'*}?key=${key%%$'\n'*}"
./hosts/up_axfc.sh:207: elif grep -Eqi 'Axfc Uploader -投稿完了.*キーワード付きURL.*a href="https://www.axfc.net.*(QueryString無しVer)' <<< "${response}" ; then
./hosts/up_bedrive.sh:99: PostUrlHost='https://bedrive.ru/core/page/ajax/file_upload_handler.ajax.php?r=bedrive.ru'
./hosts/up_bedrive.sh:110: if grep -Eqi '"error":null,"url":"https:\\/\\/bedrive.ru\\/' <<< "${response}" ; then
./hosts/up_bedrive.sh:111: hash=$(grep -oPi '(?<="url":"https:\\/\\/bedrive.ru\\/).*?(?=".*$)' <<< "$response")
@ -399,75 +403,75 @@ _________________________________________________________________________
./hosts/up_uploadraja.sh:99: PostUrlHost='https://awsaisiaposisition69.kalpstudio.xyz/cgi-bin/upload.cgi?upload_type=file&utype=anon'
./hosts/up_uploadraja.sh:119: downloadLink="https://uploadraja.com/$hash"
./hosts/up_yolobit.sh:99: PostUrlHost='https://ns08.zipcluster.com/upload.php'
./mad.sh:665: sed -i -e "s>^${url/https:/http:}.*>#& #OK# ${filename}>g" "${InputFile}" #http (if changed)
./mad.sh:667: sed -i -e "s>^direct=${url/https:/http:}.*>#& #OK# ${filename}>g" "${InputFile}" #direct url https
./mad.sh:670: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:672: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:693: sed -i -e "s>^${url/https:/http:}.*>#& #OK# ${filename} (File exists)>g" "${InputFile}" #http (if changed)
./mad.sh:695: sed -i -e "s>^direct=${url/https:/http:}.*>#& #OK# ${filename} (File exists)>g" "${InputFile}" #direct url https
./mad.sh:698: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:700: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:721: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${message}>g" "${InputFile}" #http (if changed)
./mad.sh:723: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${message}>g" "${InputFile}" #direct url https
./mad.sh:726: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:728: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:750: sed -i -e "s>^${url/https:/http:}.*>#& #FAIL# $message>g" "${InputFile}" #http (if changed)
./mad.sh:752: sed -i -e "s>^direct=${url/https:/http:}.*>#& #FAIL# $message>g" "${InputFile}" #direct url https
./mad.sh:755: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:757: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:778: sed -i -e "s>^${url/https:/http:}.*>#& #REMOVED#>g" "${InputFile}" #http (if changed)
./mad.sh:780: sed -i -e "s>^direct=${url/https:/http:}.*>#& #REMOVED#>g" "${InputFile}" #direct url https
./mad.sh:783: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:785: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:811: sed -i -e "s>^${url/https:/http:}.*>${url}|${newfilename}>g" "${InputFile}" #http (if changed)
./mad.sh:813: sed -i -e "s>^direct=${url/https:/http:}.*>direct=${url}|${newfilename}>g" "${InputFile}" #direct url https
./mad.sh:833: sed -i -e "s%^${url/https:/http:}.*%${newurl//[[:space:]]/$'\\\n'}%g" "${InputFile}" #http (if changed)
./mad.sh:854: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (dropped or bad size)>g" "${InputFile}" #http (if changed)
./mad.sh:856: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (dropped or bad size)>g" "${InputFile}" #direct url https
./mad.sh:859: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:861: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:877: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# (No CDN found)>g" "${InputFile}" #http (if changed)
./mad.sh:879: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# (No CDN found)>g" "${InputFile}" #direct url https
./mad.sh:882: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:884: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:903: sed -i -e "s>^${url/https:/http:}.*>#& #PASSWORD#>g" "${InputFile}" #http (if changed)
./mad.sh:905: sed -i -e "s>^direct=${url/https:/http:}.*>#& #PASSWORD#>g" "${InputFile}" #direct url https
./mad.sh:908: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:910: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:930: sed -i -e "s>^${url/https:/http:}.*>#& #BAD-URL#${message}>g" "${InputFile}" #http (if changed)
./mad.sh:932: sed -i -e "s>^direct=${url/https:/http:}.*>#& #BAD-URL#${message}>g" "${InputFile}" #direct url https
./mad.sh:935: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:937: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:955: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (No filename or size. Try again later)>g" "${InputFile}" #http (if changed)
./mad.sh:957: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (No filename or size. Try again later)>g" "${InputFile}" #direct url https
./mad.sh:960: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:962: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:981: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (File exists and is too large--retry or delete from downloads)>g" "${InputFile}" #http (if changed)
./mad.sh:983: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (File exists and is too large--retry or delete from downloads)>g" "${InputFile}" #direct url https
./mad.sh:986: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:988: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:1406: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest)
./mad.sh:1423: download_url='https://github.com/lwthiker/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz'
./mad.sh:1529: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest)
./mad.sh:1546: download_url='https://github.com/lexiforest/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz'
./mad.sh:1809: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1837: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1859: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:3175: if grep -Eqi '.onion' <<< "$download_url" && grep -Eqi 'https://' <<< "$download_url" ; then
./mad.sh:3688:arg2="$2" # auto, filelist, <https://url>
./mad.sh:3785: echo -e " - http://oshi.at/abcd/origAABB.rar|My specified file.part1.rar"
./mad.sh:3787: echo -e " - direct=http://pomf2.lain.la/f/abcd00zz.7z"
./mad.sh:3789: echo -e ' - ie. direct=http://somehost.onion/abcD|filename.part1.rar'
./mad.sh:4008: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4009: remote_url=${remote_url/http:/https:}
./mad.sh:4030: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4031: remote_url=${remote_url/http:/https:}
./mad.sh:4397: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4398: remote_url=${remote_url/http:/https:}
./mad.sh:4456: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4457: remote_url=${remote_url/http:/https:}
./mad.sh:4482: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4483: remote_url=${remote_url/http:/https:}
./mad.sh:672: sed -i -e "s>^${url/https:/http:}.*>#& #OK# ${filename}>g" "${InputFile}" #http (if changed)
./mad.sh:674: sed -i -e "s>^direct=${url/https:/http:}.*>#& #OK# ${filename}>g" "${InputFile}" #direct url https
./mad.sh:677: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:679: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:700: sed -i -e "s>^${url/https:/http:}.*>#& #OK# ${filename} (File exists)>g" "${InputFile}" #http (if changed)
./mad.sh:702: sed -i -e "s>^direct=${url/https:/http:}.*>#& #OK# ${filename} (File exists)>g" "${InputFile}" #direct url https
./mad.sh:705: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:707: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:728: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${message}>g" "${InputFile}" #http (if changed)
./mad.sh:730: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${message}>g" "${InputFile}" #direct url https
./mad.sh:733: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:735: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:757: sed -i -e "s>^${url/https:/http:}.*>#& #FAIL# $message>g" "${InputFile}" #http (if changed)
./mad.sh:759: sed -i -e "s>^direct=${url/https:/http:}.*>#& #FAIL# $message>g" "${InputFile}" #direct url https
./mad.sh:762: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:764: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:785: sed -i -e "s>^${url/https:/http:}.*>#& #REMOVED#>g" "${InputFile}" #http (if changed)
./mad.sh:787: sed -i -e "s>^direct=${url/https:/http:}.*>#& #REMOVED#>g" "${InputFile}" #direct url https
./mad.sh:790: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:792: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:818: sed -i -e "s>^${url/https:/http:}.*>${url}|${newfilename}>g" "${InputFile}" #http (if changed)
./mad.sh:820: sed -i -e "s>^direct=${url/https:/http:}.*>direct=${url}|${newfilename}>g" "${InputFile}" #direct url https
./mad.sh:840: sed -i -e "s%^${url/https:/http:}.*%${newurl//[[:space:]]/$'\\\n'}%g" "${InputFile}" #http (if changed)
./mad.sh:861: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (dropped or bad size)>g" "${InputFile}" #http (if changed)
./mad.sh:863: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (dropped or bad size)>g" "${InputFile}" #direct url https
./mad.sh:866: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:868: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:884: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# (No CDN found)>g" "${InputFile}" #http (if changed)
./mad.sh:886: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# (No CDN found)>g" "${InputFile}" #direct url https
./mad.sh:889: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:891: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:910: sed -i -e "s>^${url/https:/http:}.*>#& #PASSWORD#>g" "${InputFile}" #http (if changed)
./mad.sh:912: sed -i -e "s>^direct=${url/https:/http:}.*>#& #PASSWORD#>g" "${InputFile}" #direct url https
./mad.sh:915: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:917: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:937: sed -i -e "s>^${url/https:/http:}.*>#& #BAD-URL#${message}>g" "${InputFile}" #http (if changed)
./mad.sh:939: sed -i -e "s>^direct=${url/https:/http:}.*>#& #BAD-URL#${message}>g" "${InputFile}" #direct url https
./mad.sh:942: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:944: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:962: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (No filename or size. Try again later)>g" "${InputFile}" #http (if changed)
./mad.sh:964: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (No filename or size. Try again later)>g" "${InputFile}" #direct url https
./mad.sh:967: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:969: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:988: sed -i -e "s>^${url/https:/http:}.*>#& #RETRY# ${filename} (File exists and is too large--retry or delete from downloads)>g" "${InputFile}" #http (if changed)
./mad.sh:990: sed -i -e "s>^direct=${url/https:/http:}.*>#& #RETRY# ${filename} (File exists and is too large--retry or delete from downloads)>g" "${InputFile}" #direct url https
./mad.sh:993: sed -i -e "s>^${url/https:/http:}.*>#&>g" "${InputFile}" #http (if changed)
./mad.sh:995: sed -i -e "s>^direct=${url/https:/http:}.*>#&>g" "${InputFile}" #direct url https
./mad.sh:1413: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest)
./mad.sh:1430: download_url='https://github.com/lwthiker/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz'
./mad.sh:1536: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest)
./mad.sh:1553: download_url='https://github.com/lexiforest/curl-impersonate/releases/download/'"$latestTag"'/curl-impersonate-'"$latestTag"'.x86_64-linux-gnu.tar.gz'
./mad.sh:1816: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1844: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1866: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:3182: if grep -Eqi '.onion' <<< "$download_url" && grep -Eqi 'https://' <<< "$download_url" ; then
./mad.sh:3695:arg2="$2" # auto, filelist, <https://url>
./mad.sh:3792: echo -e " - http://oshi.at/abcd/origAABB.rar|My specified file.part1.rar"
./mad.sh:3794: echo -e " - direct=http://pomf2.lain.la/f/abcd00zz.7z"
./mad.sh:3796: echo -e ' - ie. direct=http://somehost.onion/abcD|filename.part1.rar'
./mad.sh:4015: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4016: remote_url=${remote_url/http:/https:}
./mad.sh:4037: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4038: remote_url=${remote_url/http:/https:}
./mad.sh:4404: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4405: remote_url=${remote_url/http:/https:}
./mad.sh:4463: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4464: remote_url=${remote_url/http:/https:}
./mad.sh:4489: if [[ ${remote_url} =~ ^http: ]] ; then
./mad.sh:4490: remote_url=${remote_url/http:/https:}
./plugins/pjscloud.sh:51: "https://PhantomJScloud.com/api/browser/v2/$RandomPjsKey/" & sleep 8s; kill -HUP $! 2>/dev/null)
./plugins/pjscloud.sh:59: "https://PhantomJScloud.com/api/browser/v2/$RandomPjsKey/" & sleep 8s; kill -HUP $! 2>/dev/null)

View file

@ -1,4 +1,4 @@
DateTime: 24.12.26
DateTime: 25.01.04
Files:
./hosts/1fichier.sh
@ -326,54 +326,54 @@ _________________________________________________________________________
./hosts/anonfile.sh:349: printf "\\n"
./hosts/anonfile.sh:350: echo -e "${RED}| Failed to extract download link [3].${NC}"
--
./hosts/anonfile.sh:446: file_header=$(tor_curl_request -i -s --head \
./hosts/anonfile.sh:447: --referer "${fixed_url}" \
./hosts/anonfile.sh:448: "$download_url")
./hosts/anonfile.sh:449: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/anonfile.sh:450: debugHtml "${remote_url##*/}" "anon_head$j" "download_url: ${download_url}"$'\n'"${file_header}"
./hosts/anonfile.sh:451: fi
./hosts/anonfile.sh:452: if [[ -z $file_header ]] ; then
./hosts/anonfile.sh:453: if [ $j == $maxfetchretries ] ; then
./hosts/anonfile.sh:454: rm -f "${anon_cookie_jar}";
./hosts/anonfile.sh:455: printf "\\n"
./hosts/anonfile.sh:456: echo -e "${RED}| Failed to extract file info${NC}"
./hosts/anonfile.sh:451: file_header=$(tor_curl_request -i -s --head \
./hosts/anonfile.sh:452: --referer "${fixed_url}" \
./hosts/anonfile.sh:453: "$download_url")
./hosts/anonfile.sh:454: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/anonfile.sh:455: debugHtml "${remote_url##*/}" "anon_head$j" "download_url: ${download_url}"$'\n'"${file_header}"
./hosts/anonfile.sh:456: fi
./hosts/anonfile.sh:457: if [[ -z $file_header ]] ; then
./hosts/anonfile.sh:458: if [ $j == $maxfetchretries ] ; then
./hosts/anonfile.sh:459: rm -f "${anon_cookie_jar}";
./hosts/anonfile.sh:460: printf "\\n"
./hosts/anonfile.sh:461: echo -e "${RED}| Failed to extract file info${NC}"
--
./hosts/anonfile.sh:551: tor_curl_request --insecure \
./hosts/anonfile.sh:552: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/anonfile.sh:553: -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \
./hosts/anonfile.sh:554: --referer "${fixed_url}" \
./hosts/anonfile.sh:555: "$download_url" --continue-at - --output "$file_path"
./hosts/anonfile.sh:556: else
./hosts/anonfile.sh:557: tor_curl_request --insecure \
./hosts/anonfile.sh:558: -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \
./hosts/anonfile.sh:559: --referer "${fixed_url}" \
./hosts/anonfile.sh:560: "$download_url" --continue-at - --output "$file_path"
./hosts/anonfile.sh:561: fi
./hosts/anonfile.sh:562: else
./hosts/anonfile.sh:563: if [ "${RateMonitorEnabled}" == "true" ]; then
./hosts/anonfile.sh:564: tor_curl_request --insecure \
./hosts/anonfile.sh:565: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/anonfile.sh:566: -H "User-Agent: $RandomUA" \
./hosts/anonfile.sh:567: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/anonfile.sh:568: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/anonfile.sh:569: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/anonfile.sh:570: -H "Connection: keep-alive" \
./hosts/anonfile.sh:571: -H "Cookie: lng=eng" \
./hosts/anonfile.sh:572: -H "Upgrade-Insecure-Requests: 1" \
./hosts/anonfile.sh:573: -H "Sec-Fetch-Dest: document" \
./hosts/anonfile.sh:574: -H "Sec-Fetch-Mode: navigate" \
./hosts/anonfile.sh:559: tor_curl_request --insecure \
./hosts/anonfile.sh:560: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/anonfile.sh:561: -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \
./hosts/anonfile.sh:562: --referer "${fixed_url}" \
./hosts/anonfile.sh:563: "$download_url" --output "$file_path"
./hosts/anonfile.sh:564: else
./hosts/anonfile.sh:565: tor_curl_request --insecure \
./hosts/anonfile.sh:566: -b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \
./hosts/anonfile.sh:567: --referer "${fixed_url}" \
./hosts/anonfile.sh:568: "$download_url" --output "$file_path"
./hosts/anonfile.sh:569: fi
./hosts/anonfile.sh:570: else
./hosts/anonfile.sh:571: if [ "${RateMonitorEnabled}" == "true" ]; then
./hosts/anonfile.sh:572: tor_curl_request --insecure \
./hosts/anonfile.sh:573: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/anonfile.sh:574: -H "User-Agent: $RandomUA" \
./hosts/anonfile.sh:575: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/anonfile.sh:576: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/anonfile.sh:577: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/anonfile.sh:578: -H "Connection: keep-alive" \
./hosts/anonfile.sh:579: -H "Cookie: lng=eng" \
./hosts/anonfile.sh:580: -H "Upgrade-Insecure-Requests: 1" \
./hosts/anonfile.sh:581: -H "Sec-Fetch-Dest: document" \
./hosts/anonfile.sh:582: -H "Sec-Fetch-Mode: navigate" \
--
./hosts/anonfile.sh:581: tor_curl_request --insecure \
./hosts/anonfile.sh:582: -H "User-Agent: $RandomUA" \
./hosts/anonfile.sh:583: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/anonfile.sh:584: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/anonfile.sh:585: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/anonfile.sh:586: -H "Connection: keep-alive" \
./hosts/anonfile.sh:587: -H "Cookie: lng=eng" \
./hosts/anonfile.sh:588: -H "Upgrade-Insecure-Requests: 1" \
./hosts/anonfile.sh:589: -H "Sec-Fetch-Dest: document" \
./hosts/anonfile.sh:590: -H "Sec-Fetch-Mode: navigate" \
./hosts/anonfile.sh:591: -H "Sec-Fetch-Site: same-origin" \
./hosts/anonfile.sh:589: tor_curl_request --insecure \
./hosts/anonfile.sh:590: -H "User-Agent: $RandomUA" \
./hosts/anonfile.sh:591: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/anonfile.sh:592: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/anonfile.sh:593: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/anonfile.sh:594: -H "Connection: keep-alive" \
./hosts/anonfile.sh:595: -H "Cookie: lng=eng" \
./hosts/anonfile.sh:596: -H "Upgrade-Insecure-Requests: 1" \
./hosts/anonfile.sh:597: -H "Sec-Fetch-Dest: document" \
./hosts/anonfile.sh:598: -H "Sec-Fetch-Mode: navigate" \
./hosts/anonfile.sh:599: -H "Sec-Fetch-Site: same-origin" \
--
./hosts/anonsharing.sh:91: response=$(tor_curl_request --insecure -i -s \
./hosts/anonsharing.sh:92: -b "${ansh_cookie_jar}" -c "${ansh_cookie_jar}" \
@ -829,76 +829,76 @@ _________________________________________________________________________
./hosts/dashfile.sh:99: printf "\\n"
./hosts/dashfile.sh:100: echo -e "${RED}| Failed to extract download link [1]${NC}"
--
./hosts/dashfile.sh:169: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/dashfile.sh:170: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:171: --data "$form_data" "$remote_url")
./hosts/dashfile.sh:172: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/dashfile.sh:173: debugHtml "${remote_url##*/}" "dash_post1_$i" "url: ${remote_url}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
./hosts/dashfile.sh:174: fi
./hosts/dashfile.sh:175: if [[ -z $response ]] ; then
./hosts/dashfile.sh:176: if [ $i == $maxfetchretries ] ; then
./hosts/dashfile.sh:177: rm -f "${dash_cookie_jar}";
./hosts/dashfile.sh:178: printf "\\n"
./hosts/dashfile.sh:179: echo -e "${RED}| Failed to extract download link [4]${NC}"
./hosts/dashfile.sh:177: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/dashfile.sh:178: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:179: --data "$form_data" "$remote_url")
./hosts/dashfile.sh:180: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/dashfile.sh:181: debugHtml "${remote_url##*/}" "dash_post1_$i" "url: ${remote_url}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
./hosts/dashfile.sh:182: fi
./hosts/dashfile.sh:183: if [[ -z $response ]] ; then
./hosts/dashfile.sh:184: if [ $i == $maxfetchretries ] ; then
./hosts/dashfile.sh:185: rm -f "${dash_cookie_jar}";
./hosts/dashfile.sh:186: printf "\\n"
./hosts/dashfile.sh:187: echo -e "${RED}| Failed to extract download link [4]${NC}"
--
./hosts/dashfile.sh:297: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/dashfile.sh:298: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:299: --data "$form_data" "$remote_url")
./hosts/dashfile.sh:300: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/dashfile.sh:301: debugHtml "${remote_url##*/}" "dash_post2_$i" "url: ${remote_url}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
./hosts/dashfile.sh:302: fi
./hosts/dashfile.sh:303: if [[ -z $response ]] ; then
./hosts/dashfile.sh:304: if [ $i == $maxfetchretries ] ; then
./hosts/dashfile.sh:305: rm -f "${dash_cookie_jar}";
./hosts/dashfile.sh:306: printf "\\n"
./hosts/dashfile.sh:307: echo -e "${RED}| Failed to extract download link [7]${NC}"
./hosts/dashfile.sh:308: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/dashfile.sh:309: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:310: --data "$form_data" "$remote_url")
./hosts/dashfile.sh:311: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/dashfile.sh:312: debugHtml "${remote_url##*/}" "dash_post2_$i" "url: ${remote_url}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
./hosts/dashfile.sh:313: fi
./hosts/dashfile.sh:314: if [[ -z $response ]] ; then
./hosts/dashfile.sh:315: if [ $i == $maxfetchretries ] ; then
./hosts/dashfile.sh:316: rm -f "${dash_cookie_jar}";
./hosts/dashfile.sh:317: printf "\\n"
./hosts/dashfile.sh:318: echo -e "${RED}| Failed to extract download link [7]${NC}"
--
./hosts/dashfile.sh:386: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url")
./hosts/dashfile.sh:387: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/dashfile.sh:388: debugHtml "${remote_url##*/}" "dash_head$j" "download_url: ${download_url}"$'\n'"${file_header}"
./hosts/dashfile.sh:389: fi
./hosts/dashfile.sh:390: if [[ -z $file_header ]] ; then
./hosts/dashfile.sh:391: if [ $j == $maxfetchretries ] ; then
./hosts/dashfile.sh:392: rm -f "${dash_cookie_jar}";
./hosts/dashfile.sh:393: printf "\\n"
./hosts/dashfile.sh:394: echo -e "${RED}| Failed to extract file info${NC}"
./hosts/dashfile.sh:395: warnAndRetryUnknownError=true
./hosts/dashfile.sh:396: if [ "${finalAttempt}" == "true" ] ; then
./hosts/dashfile.sh:397: file_header=$(tor_curl_request --insecure -L --head -s --referer "${fixed_url}" "$download_url")
./hosts/dashfile.sh:398: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/dashfile.sh:399: debugHtml "${remote_url##*/}" "dash_head$j" "download_url: ${download_url}"$'\n'"${file_header}"
./hosts/dashfile.sh:400: fi
./hosts/dashfile.sh:401: if [[ -z $file_header ]] ; then
./hosts/dashfile.sh:402: if [ $j == $maxfetchretries ] ; then
./hosts/dashfile.sh:403: rm -f "${dash_cookie_jar}";
./hosts/dashfile.sh:404: printf "\\n"
./hosts/dashfile.sh:405: echo -e "${RED}| Failed to extract file info${NC}"
./hosts/dashfile.sh:406: warnAndRetryUnknownError=true
./hosts/dashfile.sh:407: if [ "${finalAttempt}" == "true" ] ; then
--
./hosts/dashfile.sh:486: tor_curl_request --insecure \
./hosts/dashfile.sh:487: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/dashfile.sh:488: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:489: "$download_url" --continue-at - --output "$file_path"
./hosts/dashfile.sh:490: else
./hosts/dashfile.sh:491: tor_curl_request --insecure \
./hosts/dashfile.sh:492: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:493: "$download_url" --continue-at - --output "$file_path"
./hosts/dashfile.sh:494: fi
./hosts/dashfile.sh:495: else
./hosts/dashfile.sh:496: if [ "${RateMonitorEnabled}" == "true" ]; then
./hosts/dashfile.sh:497: tor_curl_request --insecure \
./hosts/dashfile.sh:498: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/dashfile.sh:499: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:500: -H "User-Agent: $RandomUA" \
./hosts/dashfile.sh:501: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/dashfile.sh:502: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/dashfile.sh:503: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/dashfile.sh:504: -H "Connection: keep-alive" \
./hosts/dashfile.sh:505: -H "Cookie: lng=eng" \
./hosts/dashfile.sh:506: -H "Upgrade-Insecure-Requests: 1" \
./hosts/dashfile.sh:507: -H "Sec-Fetch-Dest: document" \
./hosts/dashfile.sh:500: "$download_url" --continue-at - --output "$file_path"
./hosts/dashfile.sh:501: else
./hosts/dashfile.sh:502: tor_curl_request --insecure \
./hosts/dashfile.sh:503: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:504: "$download_url" --continue-at - --output "$file_path"
./hosts/dashfile.sh:505: fi
./hosts/dashfile.sh:506: else
./hosts/dashfile.sh:507: if [ "${RateMonitorEnabled}" == "true" ]; then
./hosts/dashfile.sh:508: tor_curl_request --insecure \
./hosts/dashfile.sh:509: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/dashfile.sh:510: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:511: -H "User-Agent: $RandomUA" \
./hosts/dashfile.sh:512: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/dashfile.sh:513: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/dashfile.sh:514: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/dashfile.sh:515: -H "Connection: keep-alive" \
./hosts/dashfile.sh:516: -H "Cookie: lng=eng" \
./hosts/dashfile.sh:517: -H "Upgrade-Insecure-Requests: 1" \
./hosts/dashfile.sh:518: -H "Sec-Fetch-Dest: document" \
--
./hosts/dashfile.sh:513: tor_curl_request --insecure \
./hosts/dashfile.sh:514: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:515: -H "User-Agent: $RandomUA" \
./hosts/dashfile.sh:516: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/dashfile.sh:517: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/dashfile.sh:518: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/dashfile.sh:519: -H "Connection: keep-alive" \
./hosts/dashfile.sh:520: -H "Cookie: lng=eng" \
./hosts/dashfile.sh:521: -H "Upgrade-Insecure-Requests: 1" \
./hosts/dashfile.sh:522: -H "Sec-Fetch-Dest: document" \
./hosts/dashfile.sh:523: -H "Sec-Fetch-Mode: navigate" \
./hosts/dashfile.sh:524: tor_curl_request --insecure \
./hosts/dashfile.sh:525: -b "${dash_cookie_jar}" -c "${dash_cookie_jar}" \
./hosts/dashfile.sh:526: -H "User-Agent: $RandomUA" \
./hosts/dashfile.sh:527: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/dashfile.sh:528: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/dashfile.sh:529: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/dashfile.sh:530: -H "Connection: keep-alive" \
./hosts/dashfile.sh:531: -H "Cookie: lng=eng" \
./hosts/dashfile.sh:532: -H "Upgrade-Insecure-Requests: 1" \
./hosts/dashfile.sh:533: -H "Sec-Fetch-Dest: document" \
./hosts/dashfile.sh:534: -H "Sec-Fetch-Mode: navigate" \
--
./hosts/dataupload.sh:90: response=$(tor_curl_request --insecure -L -s -b "${dup_cookie_jar}" -c "${dup_cookie_jar}" "$remote_url")
./hosts/dataupload.sh:91: if [ "${DebugAllEnabled}" == "true" ] ; then
@ -1147,74 +1147,74 @@ _________________________________________________________________________
./hosts/fileblade.sh:99: warnAndRetryUnknownError=true
./hosts/fileblade.sh:100: if [ "${finalAttempt}" == "true" ] ; then
--
./hosts/fileblade.sh:167: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/fileblade.sh:168: -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \
./hosts/fileblade.sh:169: --data "$form_data" "$post_action")
./hosts/fileblade.sh:170: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/fileblade.sh:171: debugHtml "${remote_url##*/}" "fb_post" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
./hosts/fileblade.sh:172: fi
./hosts/fileblade.sh:173: if [[ -z $response ]] ; then
./hosts/fileblade.sh:174: if [ $i == $maxfetchretries ] ; then
./hosts/fileblade.sh:175: rm -f "${fb_cookie_jar}";
./hosts/fileblade.sh:176: printf "\\n"
./hosts/fileblade.sh:177: echo -e "${RED}| Failed to extract download link [3]${NC}"
./hosts/fileblade.sh:164: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/fileblade.sh:165: -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \
./hosts/fileblade.sh:166: --data "$form_data" "$post_action")
./hosts/fileblade.sh:167: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/fileblade.sh:168: debugHtml "${remote_url##*/}" "fb_post(1)" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
./hosts/fileblade.sh:169: fi
./hosts/fileblade.sh:170: if [[ -z $response ]] ; then
./hosts/fileblade.sh:171: if [ $i == $maxfetchretries ] ; then
./hosts/fileblade.sh:172: rm -f "${fb_cookie_jar}";
./hosts/fileblade.sh:173: printf "\\n"
./hosts/fileblade.sh:174: echo -e "${RED}| Failed to extract download link [3]${NC}"
--
./hosts/fileblade.sh:254: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/fileblade.sh:255: -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \
./hosts/fileblade.sh:256: --data "$form_data" "$post_action")
./hosts/fileblade.sh:257: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/fileblade.sh:258: debugHtml "${remote_url##*/}" "fb_post2" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
./hosts/fileblade.sh:259: fi
./hosts/fileblade.sh:260: if [[ -z $response ]] ; then
./hosts/fileblade.sh:261: if [ $i == $maxfetchretries ] ; then
./hosts/fileblade.sh:262: rm -f "${fb_cookie_jar}";
./hosts/fileblade.sh:263: printf "\\n"
./hosts/fileblade.sh:264: echo -e "${RED}| Failed to extract download link [4].${NC}"
./hosts/fileblade.sh:266: response=$(tor_curl_request --insecure -L -s -X POST \
./hosts/fileblade.sh:267: -b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \
./hosts/fileblade.sh:268: --data "$form_data" "$post_action")
./hosts/fileblade.sh:269: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/fileblade.sh:270: debugHtml "${remote_url##*/}" "fb_post(2)" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
./hosts/fileblade.sh:271: fi
./hosts/fileblade.sh:272: if [[ -z $response ]] ; then
./hosts/fileblade.sh:273: if [ $i == $maxfetchretries ] ; then
./hosts/fileblade.sh:274: rm -f "${fb_cookie_jar}";
./hosts/fileblade.sh:275: printf "\\n"
./hosts/fileblade.sh:276: echo -e "${RED}| Failed to extract download link [4].${NC}"
--
./hosts/fileblade.sh:323: file_header=$(tor_curl_request --insecure -L --head -s "$download_url")
./hosts/fileblade.sh:324: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/fileblade.sh:325: debugHtml "${remote_url##*/}" "fb_head$j" "download_url: ${download_url}"$'\n'"${file_header}"
./hosts/fileblade.sh:326: fi
./hosts/fileblade.sh:327: if [[ -z $file_header ]] ; then
./hosts/fileblade.sh:328: if [ $j == $maxfetchretries ] ; then
./hosts/fileblade.sh:329: rm -f "${fb_cookie_jar}";
./hosts/fileblade.sh:330: printf "\\n"
./hosts/fileblade.sh:331: echo -e "${RED}| Failed to extract file info [1]${NC}"
./hosts/fileblade.sh:332: warnAndRetryUnknownError=true
./hosts/fileblade.sh:333: if [ "${finalAttempt}" == "true" ] ; then
./hosts/fileblade.sh:335: file_header=$(tor_curl_request --insecure -L --head -s "$download_url")
./hosts/fileblade.sh:336: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/fileblade.sh:337: debugHtml "${remote_url##*/}" "fb_head$j" "download_url: ${download_url}"$'\n'"${file_header}"
./hosts/fileblade.sh:338: fi
./hosts/fileblade.sh:339: if [[ -z $file_header ]] ; then
./hosts/fileblade.sh:340: if [ $j == $maxfetchretries ] ; then
./hosts/fileblade.sh:341: rm -f "${fb_cookie_jar}";
./hosts/fileblade.sh:342: printf "\\n"
./hosts/fileblade.sh:343: echo -e "${RED}| Failed to extract file info [1]${NC}"
./hosts/fileblade.sh:344: warnAndRetryUnknownError=true
./hosts/fileblade.sh:345: if [ "${finalAttempt}" == "true" ] ; then
--
./hosts/fileblade.sh:440: tor_curl_request --insecure -L \
./hosts/fileblade.sh:441: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/fileblade.sh:442: "$download_url" --continue-at - --output "$file_path"
./hosts/fileblade.sh:443: else
./hosts/fileblade.sh:444: tor_curl_request --insecure -L \
./hosts/fileblade.sh:445: "$download_url" --continue-at - --output "$file_path"
./hosts/fileblade.sh:446: fi
./hosts/fileblade.sh:447: else
./hosts/fileblade.sh:448: if [ "${RateMonitorEnabled}" == "true" ]; then
./hosts/fileblade.sh:449: tor_curl_request --insecure \
./hosts/fileblade.sh:450: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/fileblade.sh:451: -H "User-Agent: $RandomUA" \
./hosts/fileblade.sh:452: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/fileblade.sh:453: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/fileblade.sh:454: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/fileblade.sh:455: -H "Connection: keep-alive" \
./hosts/fileblade.sh:456: -H "Cookie: lng=eng" \
./hosts/fileblade.sh:457: -H "Upgrade-Insecure-Requests: 1" \
./hosts/fileblade.sh:458: -H "Sec-Fetch-Dest: document" \
./hosts/fileblade.sh:459: -H "Sec-Fetch-Mode: navigate" \
./hosts/fileblade.sh:452: tor_curl_request --insecure -L \
./hosts/fileblade.sh:453: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/fileblade.sh:454: "$download_url" --continue-at - --output "$file_path"
./hosts/fileblade.sh:455: else
./hosts/fileblade.sh:456: tor_curl_request --insecure -L \
./hosts/fileblade.sh:457: "$download_url" --continue-at - --output "$file_path"
./hosts/fileblade.sh:458: fi
./hosts/fileblade.sh:459: else
./hosts/fileblade.sh:460: if [ "${RateMonitorEnabled}" == "true" ]; then
./hosts/fileblade.sh:461: tor_curl_request --insecure \
./hosts/fileblade.sh:462: --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/fileblade.sh:463: -H "User-Agent: $RandomUA" \
./hosts/fileblade.sh:464: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/fileblade.sh:465: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/fileblade.sh:466: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/fileblade.sh:467: -H "Connection: keep-alive" \
./hosts/fileblade.sh:468: -H "Cookie: lng=eng" \
./hosts/fileblade.sh:469: -H "Upgrade-Insecure-Requests: 1" \
./hosts/fileblade.sh:470: -H "Sec-Fetch-Dest: document" \
./hosts/fileblade.sh:471: -H "Sec-Fetch-Mode: navigate" \
--
./hosts/fileblade.sh:464: tor_curl_request --insecure \
./hosts/fileblade.sh:465: -H "User-Agent: $RandomUA" \
./hosts/fileblade.sh:466: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/fileblade.sh:467: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/fileblade.sh:468: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/fileblade.sh:469: -H "Connection: keep-alive" \
./hosts/fileblade.sh:470: -H "Cookie: lng=eng" \
./hosts/fileblade.sh:471: -H "Upgrade-Insecure-Requests: 1" \
./hosts/fileblade.sh:472: -H "Sec-Fetch-Dest: document" \
./hosts/fileblade.sh:473: -H "Sec-Fetch-Mode: navigate" \
./hosts/fileblade.sh:474: -H "Sec-Fetch-Site: same-origin" \
./hosts/fileblade.sh:476: tor_curl_request --insecure \
./hosts/fileblade.sh:477: -H "User-Agent: $RandomUA" \
./hosts/fileblade.sh:478: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/fileblade.sh:479: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/fileblade.sh:480: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/fileblade.sh:481: -H "Connection: keep-alive" \
./hosts/fileblade.sh:482: -H "Cookie: lng=eng" \
./hosts/fileblade.sh:483: -H "Upgrade-Insecure-Requests: 1" \
./hosts/fileblade.sh:484: -H "Sec-Fetch-Dest: document" \
./hosts/fileblade.sh:485: -H "Sec-Fetch-Mode: navigate" \
./hosts/fileblade.sh:486: -H "Sec-Fetch-Site: same-origin" \
--
./hosts/fileditch.sh:85: file_header=$(tor_curl_request --insecure --head -L -s "$download_url")
./hosts/fileditch.sh:86: if [ "${DebugAllEnabled}" == "true" ] ; then
@ -1404,41 +1404,41 @@ _________________________________________________________________________
./hosts/gofile.sh:179: -H "Sec-Fetch-Mode: navigate" \
./hosts/gofile.sh:180: -H "Sec-Fetch-Site: none" \
--
./hosts/gofile.sh:250: file_header=$(tor_curl_request --insecure -L --head -s \
./hosts/gofile.sh:251: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \
./hosts/gofile.sh:252: -H "User-Agent: $RandomUA" \
./hosts/gofile.sh:253: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/gofile.sh:254: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/gofile.sh:255: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/gofile.sh:256: -H "Connection: keep-alive" \
./hosts/gofile.sh:257: -H "Upgrade-Insecure-Requests: 1" \
./hosts/gofile.sh:258: -H "Sec-Fetch-Dest: document" \
./hosts/gofile.sh:259: -H "Sec-Fetch-Mode: navigate" \
./hosts/gofile.sh:260: -H "Sec-Fetch-Site: none" \
./hosts/gofile.sh:258: file_header=$(tor_curl_request --insecure -L --head -s \
./hosts/gofile.sh:259: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \
./hosts/gofile.sh:260: -H "User-Agent: $RandomUA" \
./hosts/gofile.sh:261: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/gofile.sh:262: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/gofile.sh:263: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/gofile.sh:264: -H "Connection: keep-alive" \
./hosts/gofile.sh:265: -H "Upgrade-Insecure-Requests: 1" \
./hosts/gofile.sh:266: -H "Sec-Fetch-Dest: document" \
./hosts/gofile.sh:267: -H "Sec-Fetch-Mode: navigate" \
./hosts/gofile.sh:268: -H "Sec-Fetch-Site: none" \
--
./hosts/gofile.sh:369: tor_curl_request --insecure -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/gofile.sh:370: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \
./hosts/gofile.sh:371: -H "User-Agent: $RandomUA" \
./hosts/gofile.sh:372: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/gofile.sh:373: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/gofile.sh:374: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/gofile.sh:375: -H "Connection: keep-alive" \
./hosts/gofile.sh:376: -H "Upgrade-Insecure-Requests: 1" \
./hosts/gofile.sh:377: -H "Sec-Fetch-Dest: document" \
./hosts/gofile.sh:378: -H "Sec-Fetch-Mode: navigate" \
./hosts/gofile.sh:379: -H "Sec-Fetch-Site: none" \
./hosts/gofile.sh:377: tor_curl_request --insecure -G --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
./hosts/gofile.sh:378: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \
./hosts/gofile.sh:379: -H "User-Agent: $RandomUA" \
./hosts/gofile.sh:380: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/gofile.sh:381: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/gofile.sh:382: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/gofile.sh:383: -H "Connection: keep-alive" \
./hosts/gofile.sh:384: -H "Upgrade-Insecure-Requests: 1" \
./hosts/gofile.sh:385: -H "Sec-Fetch-Dest: document" \
./hosts/gofile.sh:386: -H "Sec-Fetch-Mode: navigate" \
./hosts/gofile.sh:387: -H "Sec-Fetch-Site: none" \
--
./hosts/gofile.sh:383: tor_curl_request --insecure -G \
./hosts/gofile.sh:384: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \
./hosts/gofile.sh:385: -H "User-Agent: $RandomUA" \
./hosts/gofile.sh:386: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/gofile.sh:387: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/gofile.sh:388: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/gofile.sh:389: -H "Connection: keep-alive" \
./hosts/gofile.sh:390: -H "Upgrade-Insecure-Requests: 1" \
./hosts/gofile.sh:391: -H "Sec-Fetch-Dest: document" \
./hosts/gofile.sh:392: -H "Sec-Fetch-Mode: navigate" \
./hosts/gofile.sh:393: -H "Sec-Fetch-Site: none" \
./hosts/gofile.sh:391: tor_curl_request --insecure -G \
./hosts/gofile.sh:392: -b "${gofile_cookie_jar}" -c "${gofile_cookie_jar}" \
./hosts/gofile.sh:393: -H "User-Agent: $RandomUA" \
./hosts/gofile.sh:394: -H "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8" \
./hosts/gofile.sh:395: -H "Accept-Language: en-US,en;q=0.5" \
./hosts/gofile.sh:396: -H "Accept-Encoding: gzip, deflate, br" \
./hosts/gofile.sh:397: -H "Connection: keep-alive" \
./hosts/gofile.sh:398: -H "Upgrade-Insecure-Requests: 1" \
./hosts/gofile.sh:399: -H "Sec-Fetch-Dest: document" \
./hosts/gofile.sh:400: -H "Sec-Fetch-Mode: navigate" \
./hosts/gofile.sh:401: -H "Sec-Fetch-Site: none" \
--
./hosts/hexload.sh:108: response=$(tor_curl_request --insecure -s --data "$form_data" "https://hexload.com/download")
./hosts/hexload.sh:109: if [ "${DebugAllEnabled}" == "true" ] ; then
@ -2511,8 +2511,8 @@ _________________________________________________________________________
./hosts/up_axfc.sh:190: if [ "${DebugAllEnabled}" == "true" ] ; then
./hosts/up_axfc.sh:191: debugHtml "${filepath##*/}" "${_hostCode}_upload" "post_url: ${PostUrlHost}"$'\n'"${response}"
./hosts/up_axfc.sh:192: fi
./hosts/up_axfc.sh:193: if grep -Eqi 'Axfc Uploader -投稿完了.*キーワード付きURL.*a href="https://www.axfc.net.*(QueryString無しVer)' <<< "${response}" ; then
./hosts/up_axfc.sh:194: subSearch=$(awk '/Axfc Uploader -投稿完了/,/(QueryString無しVer)/' <<< "$response")
./hosts/up_axfc.sh:193: response_ascii=$(echo "$response" | iconv -c -f UTF-8 -t ASCII//TRANSLIT)
./hosts/up_axfc.sh:194: if [ "${DebugAllEnabled}" == "true" ] ; then
--
./hosts/up_bedrive.sh:102: response=$(tor_curl_upload --insecure -i \
./hosts/up_bedrive.sh:103: -H "Content-Type: multipart/form-data" \
@ -3068,235 +3068,235 @@ _________________________________________________________________________
./hosts/youdbox.sh:287: containsHtml=true
./hosts/youdbox.sh:288: fi
--
./mad.sh:391:tor_curl_request() {
./mad.sh:392: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:393: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:394: else
./mad.sh:395: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:396: fi
./mad.sh:397:}
./mad.sh:398:tor_curl_request_extended() {
./mad.sh:399: randomtimeout=$((30 + RANDOM % (60 - 30)))
./mad.sh:400: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:401: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@"
./mad.sh:402: else
./mad.sh:403: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@"
./mad.sh:404: fi
./mad.sh:405:}
./mad.sh:406:tor_curl_upload() {
./mad.sh:398:tor_curl_request() {
./mad.sh:399: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:400: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:401: else
./mad.sh:402: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:403: fi
./mad.sh:404:}
./mad.sh:405:tor_curl_request_extended() {
./mad.sh:406: randomtimeout=$((30 + RANDOM % (60 - 30)))
./mad.sh:407: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:408: if [ "${RateMonitorEnabled}" == "true" ]; then
./mad.sh:409: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval --compressed --globoff "$@"
./mad.sh:410: else
./mad.sh:411: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:412: fi
./mad.sh:413: else
./mad.sh:414: if [ "${RateMonitorEnabled}" == "true" ]; then
./mad.sh:415: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@"
./mad.sh:416: else
./mad.sh:408: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@"
./mad.sh:409: else
./mad.sh:410: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout $randomtimeout --compressed --globoff "$@"
./mad.sh:411: fi
./mad.sh:412:}
./mad.sh:413:tor_curl_upload() {
./mad.sh:414: if [ "${UseTorCurlImpersonate}" == "true" ]; then
./mad.sh:415: if [ "${RateMonitorEnabled}" == "true" ]; then
./mad.sh:416: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval --compressed --globoff "$@"
./mad.sh:417: else
./mad.sh:418: "${curl_impersonate[@]}" --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --compressed --globoff "$@"
./mad.sh:419: fi
./mad.sh:420: else
./mad.sh:421: if [ "${RateMonitorEnabled}" == "true" ]; then
./mad.sh:422: curl --proxy "socks5h://${tor_identity}@${TorIp}:${torPort}" --connect-timeout ${ConnectTimeout} --speed-limit $UploadSpeedMin --speed-time $UploadTimeoutInterval -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:109.0) Gecko/20100101 Firefox/115.0' -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,*/*;q=0.8' -H 'Accept-Language: en-US,en;q=0.5' -H 'Accept-Encoding: gzip, deflate, br' -H 'Origin: null' -H 'Connection: keep-alive' -H 'Upgrade-Insecure-Requests: 1' -H 'Sec-Fetch-Dest: document' -H 'Sec-Fetch-Mode: navigate' -H 'Sec-Fetch-Site: cross-site' -H 'Sec-Fetch-User: ?1' -H 'TE: trailers' --compressed --globoff "$@"
./mad.sh:423: else
--
./mad.sh:1406: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest)
./mad.sh:1407: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:1408: debugHtml "github" "lbf_inst_curlimp$j" "$response"
./mad.sh:1409: fi
./mad.sh:1410: if [ ! -z "$response" ]; then
./mad.sh:1411: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response")
./mad.sh:1412: latestBinaryDate=$(grep -oPi -m 1 '(?<=<relative-time class="no-wrap" prefix="" datetime=").*?(?=T)' <<< "$response")
./mad.sh:1413: break
./mad.sh:1414: fi
./mad.sh:1415: done
./mad.sh:1416: if [ -z $latestTag ]; then
./mad.sh:1413: response=$(tor_curl_request --insecure -L -s https://github.com/lwthiker/curl-impersonate/releases/latest)
./mad.sh:1414: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:1415: debugHtml "github" "lbf_inst_curlimp$j" "$response"
./mad.sh:1416: fi
./mad.sh:1417: if [ ! -z "$response" ]; then
./mad.sh:1418: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response")
./mad.sh:1419: latestBinaryDate=$(grep -oPi -m 1 '(?<=<relative-time class="no-wrap" prefix="" datetime=").*?(?=T)' <<< "$response")
./mad.sh:1420: break
./mad.sh:1421: fi
./mad.sh:1422: done
./mad.sh:1423: if [ -z $latestTag ]; then
--
./mad.sh:1426: file_header=$(tor_curl_request --insecure --head -Ls "$download_url")
./mad.sh:1427: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:1428: debugHtml "github" "head_inst_curlimp$j" "${file_header}"
./mad.sh:1429: fi
./mad.sh:1430: if ! grep -Eqi 'HTTP/2 200|HTTP/1.1 200|200 OK' <<< $file_header ; then
./mad.sh:1431: if ((j == 8)) ; then
./mad.sh:1432: return 1
./mad.sh:1433: else
./mad.sh:1434: continue
./mad.sh:1435: fi
./mad.sh:1436: fi
./mad.sh:1433: file_header=$(tor_curl_request --insecure --head -Ls "$download_url")
./mad.sh:1434: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:1435: debugHtml "github" "head_inst_curlimp$j" "${file_header}"
./mad.sh:1436: fi
./mad.sh:1437: if ! grep -Eqi 'HTTP/2 200|HTTP/1.1 200|200 OK' <<< $file_header ; then
./mad.sh:1438: if ((j == 8)) ; then
./mad.sh:1439: return 1
./mad.sh:1440: else
./mad.sh:1441: continue
./mad.sh:1442: fi
./mad.sh:1443: fi
--
./mad.sh:1476: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path"
./mad.sh:1477: received_file_size=0
./mad.sh:1478: if [ -f "$file_path" ] ; then
./mad.sh:1479: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]')
./mad.sh:1480: fi
./mad.sh:1481: if ((received_file_size == file_size_bytes)) ; then
./mad.sh:1482: break
./mad.sh:1483: elif ((received_file_size < file_size_bytes)) ; then
./mad.sh:1484: if ((j >= MaxDownloadRetries)) ; then
./mad.sh:1485: echo -e "${RED}| FAILED: Size mismatch after downloading${NC}"
./mad.sh:1486: exit 1
./mad.sh:1483: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path"
./mad.sh:1484: received_file_size=0
./mad.sh:1485: if [ -f "$file_path" ] ; then
./mad.sh:1486: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]')
./mad.sh:1487: fi
./mad.sh:1488: if ((received_file_size == file_size_bytes)) ; then
./mad.sh:1489: break
./mad.sh:1490: elif ((received_file_size < file_size_bytes)) ; then
./mad.sh:1491: if ((j >= MaxDownloadRetries)) ; then
./mad.sh:1492: echo -e "${RED}| FAILED: Size mismatch after downloading${NC}"
./mad.sh:1493: exit 1
--
./mad.sh:1529: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest)
./mad.sh:1530: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:1531: debugHtml "github" "lbf_inst_curlimp$j" "$response"
./mad.sh:1532: fi
./mad.sh:1533: if [ ! -z "$response" ]; then
./mad.sh:1534: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response")
./mad.sh:1535: latestBinaryDate=$(grep -oPi -m 1 '(?<=<relative-time class="no-wrap" prefix="" datetime=").*?(?=T)' <<< "$response")
./mad.sh:1536: break
./mad.sh:1537: fi
./mad.sh:1538: done
./mad.sh:1539: if [ -z $latestTag ]; then
./mad.sh:1536: response=$(tor_curl_request --insecure -L -s https://github.com/lexiforest/curl-impersonate/releases/latest)
./mad.sh:1537: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:1538: debugHtml "github" "lbf_inst_curlimp$j" "$response"
./mad.sh:1539: fi
./mad.sh:1540: if [ ! -z "$response" ]; then
./mad.sh:1541: latestTag=$(grep -oPi -m 1 '(?<=/curl-impersonate/releases/tag/).*?(?=")' <<< "$response")
./mad.sh:1542: latestBinaryDate=$(grep -oPi -m 1 '(?<=<relative-time class="no-wrap" prefix="" datetime=").*?(?=T)' <<< "$response")
./mad.sh:1543: break
./mad.sh:1544: fi
./mad.sh:1545: done
./mad.sh:1546: if [ -z $latestTag ]; then
--
./mad.sh:1549: file_header=$(tor_curl_request --insecure --head -Ls "$download_url")
./mad.sh:1550: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:1551: debugHtml "github" "head_inst_curlimp$j" "${file_header}"
./mad.sh:1552: fi
./mad.sh:1553: if ! grep -Eqi 'HTTP/2 200|HTTP/1.1 200|200 OK' <<< $file_header ; then
./mad.sh:1554: if ((j == 8)) ; then
./mad.sh:1555: return 1
./mad.sh:1556: else
./mad.sh:1557: continue
./mad.sh:1558: fi
./mad.sh:1559: fi
./mad.sh:1556: file_header=$(tor_curl_request --insecure --head -Ls "$download_url")
./mad.sh:1557: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:1558: debugHtml "github" "head_inst_curlimp$j" "${file_header}"
./mad.sh:1559: fi
./mad.sh:1560: if ! grep -Eqi 'HTTP/2 200|HTTP/1.1 200|200 OK' <<< $file_header ; then
./mad.sh:1561: if ((j == 8)) ; then
./mad.sh:1562: return 1
./mad.sh:1563: else
./mad.sh:1564: continue
./mad.sh:1565: fi
./mad.sh:1566: fi
--
./mad.sh:1599: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path"
./mad.sh:1600: received_file_size=0
./mad.sh:1601: if [ -f "$file_path" ] ; then
./mad.sh:1602: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]')
./mad.sh:1603: fi
./mad.sh:1604: if ((received_file_size == file_size_bytes)) ; then
./mad.sh:1605: break
./mad.sh:1606: elif ((received_file_size < file_size_bytes)) ; then
./mad.sh:1607: if ((j >= MaxDownloadRetries)) ; then
./mad.sh:1608: echo -e "${RED}| FAILED: Size mismatch after downloading${NC}"
./mad.sh:1609: exit 1
./mad.sh:1606: tor_curl_request --insecure -L "$download_url" --continue-at - --output "$file_path"
./mad.sh:1607: received_file_size=0
./mad.sh:1608: if [ -f "$file_path" ] ; then
./mad.sh:1609: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]')
./mad.sh:1610: fi
./mad.sh:1611: if ((received_file_size == file_size_bytes)) ; then
./mad.sh:1612: break
./mad.sh:1613: elif ((received_file_size < file_size_bytes)) ; then
./mad.sh:1614: if ((j >= MaxDownloadRetries)) ; then
./mad.sh:1615: echo -e "${RED}| FAILED: Size mismatch after downloading${NC}"
./mad.sh:1616: exit 1
--
./mad.sh:1804: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1805: echo -e "Files:"
./mad.sh:1806: echo -e "${BLUE}${fil}${NC}"
./mad.sh:1807: echo -e ""
./mad.sh:1808: echo -e ""
./mad.sh:1809: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1810: echo -e "_________________________________________________________________________"
./mad.sh:1811: echo -e "$maud_http"
./mad.sh:1812: echo -e ""
./mad.sh:1813: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})"
./mad.sh:1814: echo -e "_________________________________________________________________________"
./mad.sh:1811: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1812: echo -e "Files:"
./mad.sh:1813: echo -e "${BLUE}${fil}${NC}"
./mad.sh:1814: echo -e ""
./mad.sh:1815: echo -e ""
./mad.sh:1816: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1817: echo -e "_________________________________________________________________________"
./mad.sh:1818: echo -e "$maud_http"
./mad.sh:1819: echo -e ""
./mad.sh:1820: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})"
./mad.sh:1821: echo -e "_________________________________________________________________________"
--
./mad.sh:1817: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1818: echo -e "_________________________________________________________________________"
./mad.sh:1819: echo -e "$maud_torcurl"
./mad.sh:1820: echo -e ""
./mad.sh:1821: echo -e ""
./mad.sh:1822: done
./mad.sh:1823: else
./mad.sh:1824: cd "$ScriptDir"
./mad.sh:1825: readarray -d $'' arrFiles < <(find . -name "*.sh" -printf '%p\n' | sort -Vk1)
./mad.sh:1826: cd "$WorkDir"
./mad.sh:1827: readarray -d $'' arrFiles2 < <(find . -name "*.sh" -printf '%p\n' | sort -Vk1)
./mad.sh:1824: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1825: echo -e "_________________________________________________________________________"
./mad.sh:1826: echo -e "$maud_torcurl"
./mad.sh:1827: echo -e ""
./mad.sh:1828: echo -e ""
./mad.sh:1829: done
./mad.sh:1830: else
./mad.sh:1831: cd "$ScriptDir"
./mad.sh:1832: readarray -d $'' arrFiles < <(find . -name "*.sh" -printf '%p\n' | sort -Vk1)
./mad.sh:1833: cd "$WorkDir"
./mad.sh:1834: readarray -d $'' arrFiles2 < <(find . -name "*.sh" -printf '%p\n' | sort -Vk1)
--
./mad.sh:1832: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1833: echo -e "Files:"
./mad.sh:1834: echo -e "${BLUE}${fil}${NC}"
./mad.sh:1835: echo -e ""
./mad.sh:1836: echo -e ""
./mad.sh:1837: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1838: echo -e "_________________________________________________________________________"
./mad.sh:1839: echo -e "$maud_http"
./mad.sh:1840: echo -e ""
./mad.sh:1841: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl \"${NC})"
./mad.sh:1842: echo -e "_________________________________________________________________________"
./mad.sh:1839: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1840: echo -e "Files:"
./mad.sh:1841: echo -e "${BLUE}${fil}${NC}"
./mad.sh:1842: echo -e ""
./mad.sh:1843: echo -e ""
./mad.sh:1844: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1845: echo -e "_________________________________________________________________________"
./mad.sh:1846: echo -e "$maud_http"
./mad.sh:1847: echo -e ""
./mad.sh:1848: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl \"${NC})"
./mad.sh:1849: echo -e "_________________________________________________________________________"
--
./mad.sh:1845: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1846: echo -e "_________________________________________________________________________"
./mad.sh:1847: echo -e "$maud_torcurl"
./mad.sh:1848: echo -e ""
./mad.sh:1849: done
./mad.sh:1850: for fil in "${arrFiles2[@]}";
./mad.sh:1851: do
./mad.sh:1852: maud_http=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei '(http|https):')
./mad.sh:1853: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl')
./mad.sh:1854: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1855: echo -e "Files:"
./mad.sh:1856: echo -e "${BLUE}${fil}${NC}"
./mad.sh:1857: echo -e ""
./mad.sh:1858: echo -e ""
./mad.sh:1859: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1860: echo -e "_________________________________________________________________________"
./mad.sh:1861: echo -e "$maud_http"
./mad.sh:1862: echo -e ""
./mad.sh:1863: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})"
./mad.sh:1864: echo -e "_________________________________________________________________________"
./mad.sh:1852: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1853: echo -e "_________________________________________________________________________"
./mad.sh:1854: echo -e "$maud_torcurl"
./mad.sh:1855: echo -e ""
./mad.sh:1856: done
./mad.sh:1857: for fil in "${arrFiles2[@]}";
./mad.sh:1858: do
./mad.sh:1859: maud_http=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei '(http|https):')
./mad.sh:1860: maud_curl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep --color='always' -Ei 'curl')
./mad.sh:1861: maud_torcurl=$(grep -n -vxE '[[:blank:]]*([#].*)?' $fil | grep -A 12 --color='always' -Ei 'tor_curl')
./mad.sh:1862: echo -e "Files:"
./mad.sh:1863: echo -e "${BLUE}${fil}${NC}"
./mad.sh:1864: echo -e ""
./mad.sh:1865: echo -e ""
./mad.sh:1866: echo -e "${PINK}MAD Audit of http lines:${NC} (${GREEN}grep \"http:\" or \"https:\"${NC})"
./mad.sh:1867: echo -e "_________________________________________________________________________"
./mad.sh:1868: echo -e "$maud_http"
./mad.sh:1869: echo -e ""
./mad.sh:1870: echo -e "${PINK}MAD Audit of curl:${NC} (${GREEN}grep \"curl\"${NC})"
./mad.sh:1871: echo -e "_________________________________________________________________________"
--
./mad.sh:1867: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1868: echo -e "_________________________________________________________________________"
./mad.sh:1869: echo -e "$maud_torcurl"
./mad.sh:1870: echo -e ""
./mad.sh:1871: done
./mad.sh:1872: fi
./mad.sh:1873:}
./mad.sh:1874:madStatus() {
./mad.sh:1875: local InputFile="$1"
./mad.sh:1876: if [ "$arg1" == "status" ] ; then
./mad.sh:1877: clear
./mad.sh:1874: echo -e "${PINK}MAD Audit of tor_curl (+12 lines after):${NC} (${GREEN}grep \"tor_curl\"${NC})"
./mad.sh:1875: echo -e "_________________________________________________________________________"
./mad.sh:1876: echo -e "$maud_torcurl"
./mad.sh:1877: echo -e ""
./mad.sh:1878: done
./mad.sh:1879: fi
./mad.sh:1880:}
./mad.sh:1881:madStatus() {
./mad.sh:1882: local InputFile="$1"
./mad.sh:1883: if [ "$arg1" == "status" ] ; then
./mad.sh:1884: clear
--
./mad.sh:3192: file_header=$(tor_curl_request --insecure -m 8 -s -D - -o /dev/null \
./mad.sh:3193: -H "Connection: keep-alive" \
./mad.sh:3194: -w 'EffectiveUrl=%{url_effective}' \
./mad.sh:3195: "$download_url")
./mad.sh:3196: else
./mad.sh:3197: printf "| Retrieving Head: attempt #$j"
./mad.sh:3198: rm -f "${WorkDir}/.temp/directhead"
./mad.sh:3199: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -L -s -i "$download_url" |
./mad.sh:3200: tee "${WorkDir}/.temp/directhead" &
./mad.sh:3201: sleep 6
./mad.sh:3202: [ -s "${WorkDir}/.temp/directhead" ]
./mad.sh:3203: kill $! 2>/dev/null
./mad.sh:3204: )
./mad.sh:3205: if [ ! -f "${WorkDir}/.flocks/${remote_url//[^a-zA-Z0-9]/}" ]; then
./mad.sh:3206: touch "${WorkDir}/.flocks/${remote_url//[^a-zA-Z0-9]/}"
./mad.sh:3207: fi
./mad.sh:3208: rm -f "${WorkDir}/.temp/directhead"
./mad.sh:3209: fi
./mad.sh:3199: file_header=$(tor_curl_request --insecure -m 8 -s -D - -o /dev/null \
./mad.sh:3200: -H "Connection: keep-alive" \
./mad.sh:3201: -w 'EffectiveUrl=%{url_effective}' \
./mad.sh:3202: "$download_url")
./mad.sh:3203: else
./mad.sh:3204: printf "| Retrieving Head: attempt #$j"
./mad.sh:3205: rm -f "${WorkDir}/.temp/directhead"
./mad.sh:3206: file_header=$(tor_curl_request --insecure --head -H "Connection: keep-alive" -L -s -i "$download_url" |
./mad.sh:3207: tee "${WorkDir}/.temp/directhead" &
./mad.sh:3208: sleep 6
./mad.sh:3209: [ -s "${WorkDir}/.temp/directhead" ]
./mad.sh:3210: kill $! 2>/dev/null
./mad.sh:3211: )
./mad.sh:3212: if [ ! -f "${WorkDir}/.flocks/${remote_url//[^a-zA-Z0-9]/}" ]; then
./mad.sh:3213: touch "${WorkDir}/.flocks/${remote_url//[^a-zA-Z0-9]/}"
./mad.sh:3214: fi
./mad.sh:3215: rm -f "${WorkDir}/.temp/directhead"
./mad.sh:3216: fi
--
./mad.sh:3327: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --output "$file_path"
./mad.sh:3328: rc=$?
./mad.sh:3329: if [ $rc -ne 0 ] ; then
./mad.sh:3330: printf "${RED}Download Failed (bad exit status).${NC}"
./mad.sh:3331: if [ -f ${file_path} ]; then
./mad.sh:3332: printf "${YELLOW} Partial removed...${NC}"
./mad.sh:3333: printf "\n\n"
./mad.sh:3334: rm -f "${file_path}"
./mad.sh:3335: else
./mad.sh:3336: printf "\n\n"
./mad.sh:3337: fi
./mad.sh:3334: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --output "$file_path"
./mad.sh:3335: rc=$?
./mad.sh:3336: if [ $rc -ne 0 ] ; then
./mad.sh:3337: printf "${RED}Download Failed (bad exit status).${NC}"
./mad.sh:3338: if [ -f ${file_path} ]; then
./mad.sh:3339: printf "${YELLOW} Partial removed...${NC}"
./mad.sh:3340: printf "\n\n"
./mad.sh:3341: rm -f "${file_path}"
./mad.sh:3342: else
./mad.sh:3343: printf "\n\n"
./mad.sh:3344: fi
--
./mad.sh:3371: tor_curl_request --insecure -L --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path"
./mad.sh:3372: else
./mad.sh:3373: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --continue-at - --output "$file_path"
./mad.sh:3374: fi
./mad.sh:3375: received_file_size=0
./mad.sh:3376: if [ -f "$file_path" ] ; then
./mad.sh:3377: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]')
./mad.sh:3378: fi
./mad.sh:3379: if CheckNoHtml "$remote_url" "$filename" "$file_path" ; then
./mad.sh:3380: containsHtml=false
./mad.sh:3381: else
./mad.sh:3382: containsHtml=true
./mad.sh:3383: fi
./mad.sh:3378: tor_curl_request --insecure -L --speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval "$download_url" --continue-at - --output "$file_path"
./mad.sh:3379: else
./mad.sh:3380: tor_curl_request --insecure -L --referer "$file_url" "$download_url" --continue-at - --output "$file_path"
./mad.sh:3381: fi
./mad.sh:3382: received_file_size=0
./mad.sh:3383: if [ -f "$file_path" ] ; then
./mad.sh:3384: received_file_size=$(stat --format="%s" "$file_path" | tr -d '[:space:]')
./mad.sh:3385: fi
./mad.sh:3386: if CheckNoHtml "$remote_url" "$filename" "$file_path" ; then
./mad.sh:3387: containsHtml=false
./mad.sh:3388: else
./mad.sh:3389: containsHtml=true
./mad.sh:3390: fi
--
./mad.sh:3571: response=$(tor_curl_upload --insecure -i \
./mad.sh:3572: -H "Content-Type: multipart/form-data" \
./mad.sh:3573: -F "key=" \
./mad.sh:3574: -F "time=$jira_timeval" \
./mad.sh:3575: -F "file=@${filepath}" \
./mad.sh:3576: "${jira_PostUrlHost}")
./mad.sh:3577: else
./mad.sh:3578: response=$(tor_curl_upload --insecure -i \
./mad.sh:3579: -H "Content-Type: multipart/form-data" \
./mad.sh:3580: -F "key=" \
./mad.sh:3581: -F "time=$jira_timeval" \
./mad.sh:3582: -F "files[]=@${arrFiles[@]}" \
./mad.sh:3582: -F "file=@${filepath}" \
./mad.sh:3583: "${jira_PostUrlHost}")
./mad.sh:3584: fi
./mad.sh:3585: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:3586: debugHtml "${filepath##*/}" "${_hostCode}_upload" "post_url: ${jira_PostUrlHost}"$'\n'"${response}"
./mad.sh:3587: fi
./mad.sh:3588: if grep -Eqi ' 200 ' <<< "${response}" ; then
./mad.sh:3584: else
./mad.sh:3585: response=$(tor_curl_upload --insecure -i \
./mad.sh:3586: -H "Content-Type: multipart/form-data" \
./mad.sh:3587: -F "key=" \
./mad.sh:3588: -F "time=$jira_timeval" \
./mad.sh:3589: -F "files[]=@${arrFiles[@]}" \
./mad.sh:3590: "${jira_PostUrlHost}")
./mad.sh:3591: fi
./mad.sh:3592: if [ "${DebugAllEnabled}" == "true" ] ; then
./mad.sh:3593: debugHtml "${filepath##*/}" "${_hostCode}_upload" "post_url: ${jira_PostUrlHost}"$'\n'"${response}"
./mad.sh:3594: fi
./mad.sh:3595: if grep -Eqi ' 200 ' <<< "${response}" ; then

28
hosts/anonfile.sh Normal file → Executable file
View file

@ -1,6 +1,6 @@
#! Name: anonfile.sh
#! Author: kittykat
#! Version: 2024.12.26
#! Version: 2024.12.28
#! Desc: Add support for downloading and processing of urls for a new host
#! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder
#!
@ -137,7 +137,7 @@ anon_FetchFileInfo() {
fi
if grep -Eqi 'input type="hidden" name="id" value="' <<< "$response"; then
printf "\\n"
echo -e "${GREEN}| Captcha found.${NC}"
echo -e "${GREEN}| Post link found.${NC}"
post_op=$(grep -oP '(?<=input type="hidden" name="op" value=").*(?=">)' <<< "$response")
post_id=$(grep -oP '(?<=input type="hidden" name="id" value=").*(?=">)' <<< "$response")
post_fname=$(grep -oP '(?<=input type="hidden" name="fname" value=").*(?=">)' <<< "$response")
@ -327,8 +327,8 @@ anon_FetchFileInfo() {
fi
fi
done
echo -e "| Captcha countdown (10s)…"
sleep 10s
echo -e "| Captcha countdown (5s)…"
sleep 5s
maxfetchretries=1
echo -e "${GREEN}# Fetching download url…${NC}"
for ((i=1; i<=$maxfetchretries; i++)); do
@ -413,7 +413,12 @@ anon_FetchFileInfo() {
fi
if grep -Eqi '<a class="stretched-link" href="https://anonfile.de' <<< "$response"; then
printf "\\n"
echo -e "${GREEN}| Download url found.${NC}"
echo -e "${GREEN}| Download url found [1]${NC}"
download_url=$(grep -oP '(?<=<a class="stretched-link" href=").*?(?=">.*$)' <<< "$response")
download_url=$(urlencode_literal_grouped_case_urlendingonly "$download_url")
elif grep -Eqi '<a class="stretched-link" href="' <<< "$response"; then
printf "\\n"
echo -e "${GREEN}| Download url found [2]${NC}"
download_url=$(grep -oP '(?<=<a class="stretched-link" href=").*?(?=">.*$)' <<< "$response")
download_url=$(urlencode_literal_grouped_case_urlendingonly "$download_url")
fi
@ -531,12 +536,15 @@ anon_FetchFileInfo() {
#! ----------- (3) Fetch File / Download File Function --------------- #
#!
anon_GetFile() {
echo -e "${GREEN}# Downloading…"
echo -e "${GREEN}# Downloading… ${BLUE}(No Resume)${NC}"
echo -e "${YELLOW}| File path:${NC}\t./.inflight/${filename}\n"
fileCnt=$1
retryCnt=$2
finalAttempt=$3
flockDownload="${WorkDir}/.flocks/${filename//[^a-zA-Z0-9\.\_\-]/}.flock"
if [ -f "$file_path" ]; then
rm -f "$file_path"
fi
for ((j=1; j<=$MaxDownloadRetries; j++)); do
pd_presize=0
if [ -f "$file_path" ] ; then
@ -552,12 +560,12 @@ anon_GetFile() {
--speed-limit $DownloadSpeedMin --speed-time $DownloadTimeoutInterval \
-b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \
--referer "${fixed_url}" \
"$download_url" --continue-at - --output "$file_path"
"$download_url" --output "$file_path"
else
tor_curl_request --insecure \
-b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \
--referer "${fixed_url}" \
"$download_url" --continue-at - --output "$file_path"
"$download_url" --output "$file_path"
fi
else
if [ "${RateMonitorEnabled}" == "true" ]; then
@ -576,7 +584,7 @@ anon_GetFile() {
-H "Sec-Fetch-User: ?1" \
-b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \
--referer "${fixed_url}" \
"$download_url" --continue-at - --output "$file_path"
"$download_url" --output "$file_path"
else
tor_curl_request --insecure \
-H "User-Agent: $RandomUA" \
@ -592,7 +600,7 @@ anon_GetFile() {
-H "Sec-Fetch-User: ?1" \
-b "${anon_cookie_jar}" -c "${anon_cookie_jar}" \
--referer "${fixed_url}" \
"$download_url" --continue-at - --output "$file_path"
"$download_url" --output "$file_path"
fi
fi
received_file_size=0

53
hosts/dashfile.sh Normal file → Executable file
View file

@ -1,6 +1,6 @@
#! Name: dashfile.sh
#! Author: kittykat
#! Version: 2024.12.25
#! Version: 2025.01.02
#! Desc: Add support for downloading and processing of urls for a new host
#! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder
#!
@ -107,7 +107,7 @@ dash_FetchFileInfo() {
continue
fi
fi
if grep -Eqi 'File Not Found|No such file with this filename|File was deleted|<table id="error_message"' <<< "$response"; then
if grep -Eqi 'No such file with this filename|File was deleted|<table id="error_message"' <<< "$response"; then
rm -f "${dash_cookie_jar}";
printf "\\n"
echo -e "${RED}| The file was not found. It could be deleted or expired.${NC}"
@ -115,6 +115,14 @@ dash_FetchFileInfo() {
removedDownload "${remote_url}"
return 1
fi
if grep -Eqi 'This file reached max downloads limit' <<< "$response"; then
rm -f "${dash_cookie_jar}";
printf "\\n"
echo -e "${RED}| The file has reached max downloads limit${NC}"
exitDownloadError=true
removedDownload "${remote_url}"
return 1
fi
if grep -Eqi 'name="method_free" value="Free Download">' <<< "$response"; then
printf "\\n"
echo -e "${GREEN}| Post link found.${NC}"
@ -194,7 +202,15 @@ dash_FetchFileInfo() {
exitDownloadError=true
removedDownload "${remote_url}"
return 1
fi
fi
if grep -Eqi 'This file reached max downloads limit' <<< "$response"; then
rm -f "${dash_cookie_jar}";
printf "\\n"
echo -e "${RED}| The file has reached max downloads limit${NC}"
exitDownloadError=true
removedDownload "${remote_url}"
return 1
fi
if grep -Eqi 'you have to wait|seconds till next download' <<< "$response"; then
if [ $i == $maxfetchretries ] ; then
rm -f "${dash_cookie_jar}";
@ -220,9 +236,21 @@ dash_FetchFileInfo() {
pval4=$(grep -oP -m 1 '<span style='"'"'position:absolute;padding-left:6[0-9]px;padding-top:[0-9]+px;'"'"'>&#\K.*?(?=;</span>)' <<< "$codeline" )
val1=$((pval1-48)); val2=$((pval2-48)); val3=$((pval3-48)); val4=$((pval4-48))
captcha_code="${val1}${val2}${val3}${val4}"
if grep -Eqi 'name="method_free" value="Free Download">' <<< "$response"; then
if [ -z "$captcha_code" ] || grep -Eqi '-' <<< "$captcha_code"; then
rm -f "${dash_cookie_jar}";
printf "\\n"
echo -e "${RED}| Bad captcha code [2]${NC}"
warnAndRetryUnknownError=true
if [ "${finalAttempt}" == "true" ] ; then
failedRetryDownload "${remote_url}" "Bad captcha code [2]" ""
fi
return 1
else
printf "\\n"
echo -e "${GREEN}| Captcha found.${NC}"
fi
if grep -Eqi 'name="method_free" value="Free Download">' <<< "$response"; then
echo -e "${GREEN}| Post2 found.${NC}"
post_op=$(grep -oP '(?<=input type="hidden" name="op" value=").*(?=">)' <<< "$response")
post_id=$(grep -oP '(?<=input type="hidden" name="id" value=").*(?=">)' <<< "$response")
post_rand=$(grep -oP '(?<=input type="hidden" name="rand" value=").*(?=">)' <<< "$response")
@ -231,7 +259,6 @@ dash_FetchFileInfo() {
else
if [ $i == $maxfetchretries ] ; then
rm -f "${dash_cookie_jar}";
printf "\\n"
echo -e "${RED}| Failed to extract download link [5]${NC}"
warnAndRetryUnknownError=true
if [ "${finalAttempt}" == "true" ] ; then
@ -243,25 +270,9 @@ dash_FetchFileInfo() {
continue
fi
fi
if grep -Eqi '-' <<< "$captcha_code"; then
if [ $i == $maxfetchretries ] ; then
rm -f "${dash_cookie_jar}";
printf "\\n"
echo -e "${RED}| Bad captcha code [2]${NC}"
warnAndRetryUnknownError=true
if [ "${finalAttempt}" == "true" ] ; then
failedRetryDownload "${remote_url}" "Bad captcha code [2]" ""
fi
return 1
else
tor_identity="${RANDOM}"
continue
fi
fi
if [[ -z "$captcha_code" ]] || [[ -z "$post_op" ]] || [[ -z "$post_id" ]] || [[ -z "$post_rand" ]] ; then
if [ $i == $maxfetchretries ] ; then
rm -f "${dash_cookie_jar}";
printf "\\n"
echo -e "${RED}| Failed to extract download link [6]${NC}"
warnAndRetryUnknownError=true
if [ "${finalAttempt}" == "true" ] ; then

25
hosts/eternalhosting.sh Normal file → Executable file
View file

@ -1,6 +1,6 @@
#! Name: eternalhosting.sh
#! Author: kittykat
#! Version: 2024.09.13
#! Version: 2024.12.28
#! Desc: Add support for downloading and processing of urls for a new host
#! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder
#!
@ -15,7 +15,7 @@
#! HostDomainRegex: The regex used to verify matching urls
HostCode='eh'
HostNick='eternal hosting'
HostFuncPrefix='direct'
HostFuncPrefix='eh'
HostUrls='eternalcbrzpicytj4zyguygpmkjlkddxob7tptlr25cdipe5svyqoqd.onion'
HostDomainRegex='^(http|https)://eternalcbrzpicytj4zyguygpmkjlkddxob7tptlr25cdipe5svyqoqd\.onion/file/'
#!
@ -26,6 +26,21 @@ ListHostAndDomainRegexes=${ListHostAndDomainRegexes}'/'${HostCode}'/'${HostNick}
#!
#! ------------ (1) Host Main Download Function --------------- #
#!
#! This is a direct= download host, so all the functions are already in mad.sh
#! Since the HostFuncPrefix is defined above as "direct", nothing further needs to be done as it will
#! call the direct_DownloadFile() function already in mad.sh
#! @REQUIRED: Host Main Download function
#! Must be named specifically as such:
#! <HostFuncPrefix>_DownloadFile()
eh_DownloadFile() {
local pUrl="$1"
local pFileCnt="$2"
local pUrlMod="$pUrl"
if grep -Eqi '\.onion' <<< "$pUrlMod" && grep -Eqi 'https://' <<< "$pUrlMod" ; then
echo -e "${PINK}| Reverting .onion address to http...${NC}"
pUrlMod="${pUrlMod/https/http}"
fi
if grep -Eqi '\.onion/file/' <<< "$pUrlMod" ; then
echo -e "${PINK}| Switching to download url...${NC}"
pUrlMod="${pUrlMod/\.onion\/file\//\.onion\/download\/}"
fi
echo -e "[${BLUE}ModifiedUrl${NC}]: ${pUrlMod}"
direct_DownloadFile "$pUrl" "$pFileCnt" "$pUrlMod"
}

32
hosts/fileblade.sh Normal file → Executable file
View file

@ -1,6 +1,6 @@
#! Name: isupload.sh
#! Author: kittykat
#! Version: 2024.12.20
#! Version: 2024.12.28
#! Desc: Add support for downloading and processing of urls for a new host
#! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder
#!
@ -136,7 +136,6 @@ fb_FetchFileInfo() {
post_id=$(grep -oPi -m 1 '(?<=input type="hidden" name="id" value=").*(?=">.*$)' <<< "$response")
post_fname=$(grep -oPi -m 1 '(?<=input type="hidden" name="fname" value=").*(?=">.*$)' <<< "$response")
post_referer=$(grep -oPi -m 1 '(?<=input type="hidden" name="referer" value=").*(?=">.*$)' <<< "$response")
post_action=$(urlencode_literal_grouped_case_urlendingonly "${post_action}")
fi
if [[ -z "$post_action" ]] || [[ -z "$post_op" ]] || [[ -z "$post_id" ]] || [[ -z "$post_fname" ]] ; then
rm -f "${fb_cookie_jar}";
@ -155,20 +154,18 @@ fb_FetchFileInfo() {
break
fi
done
echo -e "| Download countdown (10s)…"
sleep 10s
echo -e "${GREEN}# Fetching download2…${NC}"
for ((i=1; i<=$maxfetchretries; i++)); do
printf " _"
download_url=""
CLEANSTRING=${remote_url//[^a-zA-Z0-9]/}
trap "rm -f ${WorkDir}/.flocks/${CLEANSTRING}; echo ""; rm -f $fb_cookie_jar; tput cnorm; exit" 0 1 2 3 6 15
form_data="op=$post_op&usr_login=$post_usr_login&id=$post_id&fname=$post_fname&referer=$post_referer&method_free=method_free"
form_data="op=$post_op&usr_login=&id=$post_id&fname=$post_fname&referer=&method_free=method_free"
response=$(tor_curl_request --insecure -L -s -X POST \
-b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \
--data "$form_data" "$post_action")
if [ "${DebugAllEnabled}" == "true" ] ; then
debugHtml "${remote_url##*/}" "fb_post" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
debugHtml "${remote_url##*/}" "fb_post(1)" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
fi
if [[ -z $response ]] ; then
if [ $i == $maxfetchretries ] ; then
@ -201,14 +198,29 @@ fb_FetchFileInfo() {
failedRetryDownload "${remote_url}" "Pro download only. [Free users not allowed download over 100MB]" ""
return 1
fi
if grep -Eqi 'Just a moment...' <<< "$response"; then
if grep -Eqi 'Your subsequent download will be started in' <<< "$response"; then
if [ $i == $maxfetchretries ] ; then
rm -f "${fb_cookie_jar}";
printf "\\n"
echo -e "${RED}| Failed to extract download link [3].${NC}"
echo -e "${RED}| Subsequent download wait.. [3b]${NC}"
warnAndRetryUnknownError=true
if [ "${finalAttempt}" == "true" ] ; then
failedRetryDownload "${remote_url}" "" ""
failedRetryDownload "${remote_url}" "Subsequent download wait.. [3b]" ""
fi
return 1
else
tor_identity="${RANDOM}"
continue
fi
fi
if grep -Eqi '<p class="text-warning mb-0">' <<< "$response"; then
if [ $i == $maxfetchretries ] ; then
rm -f "${fb_cookie_jar}";
printf "\\n"
echo -e "${RED}| Failed to extract download link (Unknown warning encountered) [3c]${NC}"
warnAndRetryUnknownError=true
if [ "${finalAttempt}" == "true" ] ; then
failedRetryDownload "${remote_url}" "Unknown warning encountered in download2 [3c]" ""
fi
return 1
else
@ -255,7 +267,7 @@ fb_FetchFileInfo() {
-b "${fb_cookie_jar}" -c "${fb_cookie_jar}" \
--data "$form_data" "$post_action")
if [ "${DebugAllEnabled}" == "true" ] ; then
debugHtml "${remote_url##*/}" "fb_post2" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
debugHtml "${remote_url##*/}" "fb_post(2)" "post_action: ${post_action}"$'\n'"form_data: ${form_data}"$'\n'"${response}"
fi
if [[ -z $response ]] ; then
if [ $i == $maxfetchretries ] ; then

12
hosts/gofile.sh Normal file → Executable file
View file

@ -1,6 +1,6 @@
#! Name: gofile.sh
#! Author: kittykat
#! Version: 2024.09.13
#! Version: 2025.01.03
#! Desc: Add support for downloading and processing of urls for a new host
#! Usage: Copy this file into the ./${ScriptDir}/hosts/ folder
#!
@ -197,6 +197,14 @@ gofile_FetchFileInfo() {
else
continue
fi
fi
if grep -Eqi '"children":\{\}' <<< "$response"; then
rm -f "${gofile_cookie_jar}";
printf "\\n"
echo -e "${RED}| Bulk download is a Premium feature. (No children)${NC}"
exitDownloadError=true
removedDownload "${remote_url}" "Bulk download is a Premium feature"
return 1
fi
if grep -Eqi '"status":"ok"' <<< "$response"; then
download_url=$(grep -oPi '(?<="link":").*?(?=")' <<< "$response")
@ -225,7 +233,7 @@ gofile_FetchFileInfo() {
return 1
else
continue
fi
fi
fi
else
rm -f "${gofile_cookie_jar}";

20
hosts/up_axfc.sh Normal file → Executable file
View file

@ -1,6 +1,6 @@
#! Name: up_axfc.sh
#! Author: kittykat
#! Version: 2024.12.26
#! Version: 2025.01.02
#! Desc: Add support for uploading files to a new host
#! Info: Files are accessible at https://www.axfc.net/<hash>
#! MaxSize: 2GB
@ -190,14 +190,28 @@ axfc_PostFile() {
if [ "${DebugAllEnabled}" == "true" ] ; then
debugHtml "${filepath##*/}" "${_hostCode}_upload" "post_url: ${PostUrlHost}"$'\n'"${response}"
fi
if grep -Eqi 'Axfc Uploader -投稿完了.*キーワード付きURL.*a href="https://www.axfc.net.*(QueryString無しVer)' <<< "${response}" ; then
response_ascii=$(echo "$response" | iconv -c -f UTF-8 -t ASCII//TRANSLIT)
if [ "${DebugAllEnabled}" == "true" ] ; then
debugHtml "${filepath##*/}" "${_hostCode}_upload_ascii" "post_url: ${PostUrlHost}"$'\n'"${response_ascii}"
fi
if grep -Eqi -m 1 'a href="https://www\.axfc\.net\/u\/.*\?key=1234"' <<< "${response_ascii}" ; then
url=$(grep -oPi -m 1 '(?<=a href="https://www\.axfc\.net\/u\/).*?(?=".*$)' <<< "$response_ascii")
key=$(grep -oPi -m 1 '(?<=\?key=).*?(?=".*$)' <<< "$response_ascii")
filesize=$(GetFileSize "$filepath" "false")
downloadLink="https://www.axfc.net/u/${url%%$'\n'*}?key=${key%%$'\n'*}"
echo -e "${GREEN}| Upload Success${NC}"
echo -e "| Size: ${BLUE}${filesize}${NC} bytes${NC}"
echo -e "| Link [1]: ${YELLOW}${downloadLink}${NC}"
successUpload "$pline" "${filepath}" "${_hostCode}" "${filesize}" "${downloadLink}" "{$response}"
return 0
elif grep -Eqi 'Axfc Uploader -投稿完了.*キーワード付きURL.*a href="https://www.axfc.net.*(QueryString無しVer)' <<< "${response}" ; then
subSearch=$(awk '/Axfc Uploader -投稿完了/,/(QueryString無しVer)/' <<< "$response")
url=$(grep -oPi -m 1 '(?<=キーワード付きURL\</td\>\</tr\>\<tr\>\<td\>\<a href\=").*?(?=" target="_blank">ファイルIDのみ.*$)' <<< "$subSearch")
filesize=$(GetFileSize "$filepath" "false")
downloadLink="${url%%$'\n'*}"
echo -e "${GREEN}| Upload Success${NC}"
echo -e "| Size: ${BLUE}${filesize}${NC} bytes${NC}"
echo -e "| Link: ${YELLOW}${downloadLink}${NC}"
echo -e "| Link [2]: ${YELLOW}${downloadLink}${NC}"
successUpload "$pline" "${filepath}" "${_hostCode}" "${filesize}" "${downloadLink}" "{$response}"
return 0
else

11
mad.sh
View file

@ -1,6 +1,6 @@
#!/bin/bash
# Copyright 2023-2024 kittykat (kittykat@morke.org) | (luvherpurr@torbox3uiot6wchz.onion)
# Copyright 2023-2025 kittykat (kittykat@morke.org) | (luvherpurr@torbox3uiot6wchz.onion)
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
@ -31,9 +31,16 @@
# * klonkerz - feedback and suggestions, url only processing
# * Everyone who provided feedback and helped test.. and those who wish to remain anonymous
ScriptVersion=2024.12.26
ScriptVersion=2025.01.04
#=================================================
# Recent Additions
# 2024.01.03 - [gofile] Detect "Bulk download is a Premium feature" response (no children)
# 2025.01.02 - [up_axfc] Update PUT response check to handle kanji chars (remove)
# 2025.01.02 - [dashfile] Add response 'This file reached max downloads limit'. New cookie on captcha fail
# 2024.12.28 - [dashfile] Update captcha code check
# 2024.12.28 - [anonfile] Add new download link href response
# 2024.12.28 - [fileblade] Add additional response handling (subsequent downloads, unknown warnings)
# 2024.12.28 - [eternalhosting] Update eternal.onion to handle landing page (eternal.onion/file/)
# 2024.12.26 - [up_kouploader / up_axfc / up_torup] Fixed failedRetryUpload (was using download logging)
# * Thanks Belky
# 2024.12.26 - [anonfile / up_anonfile] Add anonfile.de as upload / download host