[ { "type": "metadata", "id": "repository_tree_map", "description": "Visual map of the fetched repository structure", "content": [ "├── .github", "│ └── workflows", "│ └── main.yml", "├── backups", "│ └── backup1", "│ └── makefolder", "├── c", "│ └── c++", "│ └── 1.cpp", "├── containers", "│ ├── docker", "│ │ └── googlecloud", "│ │ ├── cloudbuild.json", "│ │ └── Dockerfile", "│ └── vm", "│ └── makemeafolder.txt", "├── google_apps_script", "│ ├── 1.gs", "│ └── 2.gs", "├── kotlin", "│ ├── old", "│ │ └── 1old1.kt", "│ └── 1.kt", "├── old", "│ ├── hybrid_filesold0.html", "│ ├── indexold1.html", "│ ├── repo_to_jsonold1.html", "│ ├── repo_to_jsonold2.html", "│ ├── repo_to_jsonold3.html", "│ ├── website_changerold1.html", "│ ├── website_changerold2.html", "│ ├── website_changerold3.html", "│ ├── website_changerold4.html", "│ ├── website_changerold5.html", "│ └── website_changerold6.html", "├── python", "│ ├── clean_c_executer_2.py", "│ └── clean_c_executer.py", "├── scratchpad", "│ ├── game", "│ │ └── helper", "│ │ ├── dimensional_vector_optimizer.html", "│ │ └── healoptimal.html", "│ ├── hatch", "│ │ ├── primitive_dimensional_subspace.html", "│ │ ├── primitive_risky_website_inspector.html", "│ │ ├── qrreaderold0.html", "│ │ └── sw.js", "│ ├── modeltester", "│ │ └── prompts", "│ │ └── log1.txt", "│ ├── temp", "│ │ └── out", "│ │ └── html", "│ │ ├── 0079.html", "│ │ ├── communicationhelpertrversion.html", "│ │ ├── jitlogtest.html", "│ │ └── primeexp.html", "│ ├── 1.cpp", "│ └── 1.js", "├── shell", "│ ├── old", "│ │ ├── 3old1.sh", "│ │ ├── 3old2.sh", "│ │ ├── 3old3.sh", "│ │ ├── 5old1.sh", "│ │ ├── 6old1.sh", "│ │ └── 6old2.sh", "│ ├── 1.sh", "│ ├── 2.sh", "│ ├── 3.sh", "│ ├── 4.sh", "│ ├── 5.sh", "│ └── 6.sh", "├── text", "│ └── foldermake.txt", "├── things", "│ └── esp32", "│ └── sentry1.cpp", "├── 91_character_black_icon.html", "├── enhanced_qr_scanner.html", "├── enhanced_timer.html", "├── file_renamer_and_editor.html", "├── full_account_via_githack.html", "├── headphone_test.html", "├── HTML_realizer.html", "├── html_sites_only_full_account_via_hithack.html", "├── icon_tester.html", "├── index.html", "├── primitive_file_comparer.html", "├── primitive_file_list_analyser.html", "├── primitive_gemini_api_chat.html", "├── README.md", "├── repo_to_json.html", "├── Speechtotext.html", "└── website_changer.html" ] }, { "name": ".github", "type": "folder", "size": "13,140 bytes", "path": ".github", "children": [ { "name": "workflows", "type": "folder", "size": "13,140 bytes", "path": ".github/workflows", "children": [ { "name": "main.yml", "type": "file", "size": "13,140 bytes", "path": ".github/workflows/main.yml", "content": "name: url to url data transfer\npermissions:\n contents: write\n statuses: write\nconcurrency:\n group: copy-${{ github.workflow }}-${{ github.event.inputs.destination_url }}\n cancel-in-progress: false\non:\n workflow_dispatch:\n inputs:\n source_urls:\n description: 'One or more source URLs (can be /tree/ or /blob/), one per line.'\n required: true\n type: string\n source_keys:\n description: 'OPTIONAL: Your \"keychain\" of PATs, one per line. WARNING: inputs are visible in run history.'\n required: false\n type: string\n filter_mode:\n description: 'Set the filter mode.'\n required: true\n type: choice\n options: [Whitelist, Blacklist]\n default: 'Whitelist'\n filter_list:\n description: 'OPTIONAL: List of files/folders for the filter (one per line).'\n required: false\n type: string\n destination_url:\n description: 'GitHub URL for the destination folder (MUST be a /tree/ URL).'\n required: true\n type: string\n destination_key:\n description: 'OPTIONAL: PAT with push access to the destination repo/branch. Falls back to GITHUB_TOKEN (same-repo only).'\n required: false\n type: string\njobs:\n url-transfer:\n runs-on: ubuntu-latest\n outputs:\n files_copied: ${{ steps.process_sources.outputs.files_copied }}\n processed_sources: ${{ steps.process_sources.outputs.processed_sources }}\n success_count: ${{ steps.process_sources.outputs.success_count }}\n skipped_count: ${{ steps.process_sources.outputs.skipped_count }}\n no_match_count: ${{ steps.process_sources.outputs.no_match_count }}\n steps:\n - name: Validate Input URLs\n run: |\n set -euo pipefail\n DEST_URL_RAW=\"${{ github.event.inputs.destination_url }}\"\n DEST_URL=\"$(printf \"%s\" \"$DEST_URL_RAW\" | tr -d '\\r' | sed -e 's/^[[:space:]]*//' -e 's/[[:space:]]*$//')\"\n if [[ \"$DEST_URL\" != *\"/tree/\"* ]]; then echo \"::error::destination_url must be a GitHub /tree// folder URL.\"; exit 1; fi\n SRC_INPUTS_RAW=\"${{ github.event.inputs.source_urls }}\"; SRC_INPUTS=\"$(printf \"%s\" \"$SRC_INPUTS_RAW\" | tr -d '\\r')\"\n INVALID=false\n while IFS= read -r URL; do\n URL=\"$(printf \"%s\" \"$URL\" | sed -e 's/^[[:space:]]*//' -e 's/[[:space:]]*$//')\"; [[ -z \"$URL\" ]] && continue\n if [[ \"$URL\" != *\"/tree/\"* && \"$URL\" != *\"/blob/\"* ]]; then echo \"::error::Invalid source URL: $URL (must contain /tree/ or /blob/)\"; INVALID=true; fi\n done <<< \"$SRC_INPUTS\"\n if $INVALID; then echo \"ERROR: Source URL validation failed.\"; exit 1; fi\n echo \"URL validation passed.\"\n\n - name: Parse Destination URL & Set Up\n id: dest_paths\n run: |\n set -euo pipefail\n DEST_URL_RAW=\"${{ github.event.inputs.destination_url }}\"; DEST_URL=\"$(printf \"%s\" \"$DEST_URL_RAW\" | tr -d '\\r' | sed -e 's/^[[:space:]]*//' -e 's/[[:space:]]*$//')\"\n CLEAN_URL=\"$(printf \"%s\" \"$DEST_URL\" | sed -e 's|^https://||' -e 's|^http://||')\"; DEST_REPO=\"$(printf \"%s\" \"$CLEAN_URL\" | cut -d'/' -f2,3)\"\n DEST_FULL_PATH=\"$(printf \"%s\" \"$CLEAN_URL\" | sed -E 's@[^/]*/[^/]*/[^/]*/tree/@@')\"; DEST_BRANCH=\"$(printf \"%s\" \"$DEST_FULL_PATH\" | cut -d'/' -f1)\"\n DEST_FOLDER=\"$(printf \"%s\" \"$DEST_FULL_PATH\" | sed -e \"s|^$DEST_BRANCH/||\" -e \"s|^$DEST_BRANCH$||\")\"\n { echo \"DEST_REPO_NAME=$DEST_REPO\"; echo \"DEST_BRANCH=$DEST_BRANCH\"; echo \"DEST_FOLDER=$DEST_FOLDER\"; } >> \"$GITHUB_ENV\"\n\n - name: Determine Destination Token & Guard\n id: dest_token_logic\n run: |\n set -euo pipefail\n if [[ -n \"${{ github.event.inputs.destination_key }}\" ]]; then echo \"token=${{ github.event.inputs.destination_key }}\" >> \"$GITHUB_OUTPUT\"; else echo \"token=${{ secrets.GITHUB_TOKEN }}\" >> \"$GITHUB_OUTPUT\"; fi\n if [[ \"${{ steps.dest_token_logic.outputs.token }}\" == \"${{ secrets.GITHUB_TOKEN }}\" && \"${{ env.DEST_REPO_NAME }}\" != \"${{ github.repository }}\" ]]; then echo \"::error::destination_key is required for cross-repository writes.\"; exit 1; fi\n\n - name: Checkout Destination Repo (robust)\n run: |\n set -euo pipefail\n REPO=\"${{ env.DEST_REPO_NAME }}\"; BR=\"${{ env.DEST_BRANCH }}\"; TOK=\"${{ steps.dest_token_logic.outputs.token }}\"\n git clone --depth 1 \"https://${TOK}@github.com/${REPO}.git\" .\n git remote set-url origin \"https://${TOK}@github.com/${REPO}.git\"\n if git ls-remote --exit-code --heads origin \"$BR\" >/dev/null 2>&1; then git checkout -B \"$BR\" \"origin/$BR\"; else git checkout -B \"$BR\"; fi\n\n - name: Process All Sources\n id: process_sources\n run: |\n set -euo pipefail\n FILES_COPIED=false; PROCESSED_LIST=\"\"; SUCCESS_COUNT=0; SKIPPED_COUNT=0; NO_MATCH_COUNT=0\n SRC_INPUTS_RAW=\"${{ github.event.inputs.source_urls }}\"; SRC_INPUTS=\"$(printf \"%s\" \"$SRC_INPUTS_RAW\" | tr -d '\\r')\"\n FILTER_INPUTS_RAW=\"${{ github.event.inputs.filter_list }}\"; FILTER_INPUTS=\"$(printf \"%s\" \"$FILTER_INPUTS_RAW\" | tr -d '\\r')\"\n DEST_PATH=\"${{ github.workspace }}\"; if [[ -n \"${{ env.DEST_FOLDER }}\" ]]; then DEST_PATH=\"${{ github.workspace }}/${{ env.DEST_FOLDER }}\"; mkdir -p \"$DEST_PATH\"; fi\n KEYCHAIN_RAW=\"${{ github.event.inputs.source_keys }}\"; KEYCHAIN=\"$(printf \"%s\" \"$KEYCHAIN_RAW\" | tr -d '\\r')\"; mapfile -t KEYCHAIN_ARR <<< \"$KEYCHAIN\"\n resolve_branch_and_relpath() {\n local repo=\"$1\"; local full=\"$2\"; IFS='/' read -r -a PARTS <<< \"$full\"; local n=${#PARTS[@]}\n for ((i=n; i>=1; i--)); do\n local cand_branch=\"${PARTS[0]}\"; for ((k=1; k include-list.txt; : > include-effective.txt\n while IFS= read -r P; do P=\"$(printf \"%s\" \"$P\" | sed -e 's/^[[:space:]]*//' -e 's/[[:space:]]*$//')\"; [[ -z \"$P\" ]] && continue; if [[ -e \"$P\" ]]; then printf \"%s\\n\" \"$P\" >> include-effective.txt; fi; done < include-list.txt\n if [[ -s include-effective.txt ]]; then rsync -a --files-from=include-effective.txt . \"$DEST_PATH\"/; else echo \"::notice::Whitelist provided but no entries matched. Copying nothing for this source.\"; NO_MATCH_COUNT=$((NO_MATCH_COUNT + 1)); fi; rm -f include-list.txt include-effective.txt\n else\n printf \"%s\\n\" \"$FILTER_INPUTS\" | sed -e 's/^[[:space:]]*//' -e 's/[[:space:]]*$//' -e '/^$/d' > exclude-list.txt\n rsync -a --exclude-from=exclude-list.txt . \"$DEST_PATH\"/; rm -f exclude-list.txt\n fi\n fi; cd \"${{ github.workspace }}\"\n else\n if [[ -z \"$REL_PATH\" || ! -f \"$TEMP_DIR/$REL_PATH\" ]]; then echo \"::warning::file '$REL_PATH' not found in archive. Skipping.\"; SKIPPED_COUNT=$((SKIPPED_COUNT + 1)); rm -rf \"$TEMP_DIR\"; continue; fi\n (cd \"$TEMP_DIR\" && rsync -a --relative \"./$REL_PATH\" \"$DEST_PATH\"/)\n fi\n PROCESSED_LIST=\"$PROCESSED_LIST $SOURCE_REPO\"; SUCCESS_COUNT=$((SUCCESS_COUNT + 1)); rm -rf \"$TEMP_DIR\"\n done <<< \"$SRC_INPUTS\"\n if [ \"$(git status --porcelain=v1 2>/dev/null | wc -l)\" -gt 0 ]; then FILES_COPIED=true; fi\n { echo \"files_copied=$FILES_COPIED\"; echo \"processed_sources=$PROCESSED_LIST\"; echo \"success_count=$SUCCESS_COUNT\"; echo \"skipped_count=$SKIPPED_COUNT\"; echo \"no_match_count=$NO_MATCH_COUNT\"; } >> \"$GITHUB_OUTPUT\"\n\n - name: Commit and Push Changes\n if: steps.process_sources.outputs.files_copied == 'true'\n run: |\n set -euo pipefail\n git config user.name \"github-actions[bot]\"; git config user.email \"github-actions[bot]@users.noreply.github.com\"\n git fetch origin \"${{ env.DEST_BRANCH }}\" || true\n git checkout -B \"${{ env.DEST_BRANCH }}\" \"origin/${{ env.DEST_BRANCH }}\" || git checkout -B \"${{ env.DEST_BRANCH }}\"\n echo \"Rebasing before push…\"; git pull --rebase origin \"${{ env.DEST_BRANCH }}\" || echo \"No remote branch yet; continuing.\"\n git add -A; git commit -m \"Data transfer from sources:${{ steps.process_sources.outputs.processed_sources }}\" || { echo \"No changes to commit.\"; exit 0; }\n git push origin \"${{ env.DEST_BRANCH }}\"\n\n - name: Get HEAD SHA (for check-run)\n id: head\n if: always()\n run: |\n set -euo pipefail\n echo \"sha=$(git rev-parse HEAD)\" >> \"$GITHUB_OUTPUT\"\n\n - name: Publish attention status\n if: always()\n run: |\n set -euo pipefail\n SUCCESS=\"${{ steps.process_sources.outputs.success_count }}\"\n SKIPPED=\"${{ steps.process_sources.outputs.skipped_count }}\"\n NOMATCH=\"${{ steps.process_sources.outputs.no_match_count }}\"\n COPIED=\"${{ steps.process_sources.outputs.files_copied }}\"\n SUMMARY=\"Copied: ${SUCCESS:-0} | Skipped: ${SKIPPED:-0} | Whitelist no-match: ${NOMATCH:-0} | Files changed: ${COPIED:-false}\"\n STATE=\"success\"\n if [[ \"${SKIPPED:-0}\" -gt 0 || \"${SUCCESS:-0}\" -eq 0 ]]; then STATE=\"failure\"; fi\n SHA=\"${{ steps.head.outputs.sha }}\"\n if [[ -n \"$SHA\" ]]; then\n PAYLOAD=$(printf '{\"state\":\"%s\",\"description\":\"%s\",\"context\":\"Import status\"}' \"$STATE\" \"$SUMMARY\")\n curl -sL -X POST \\\n -H \"Accept: application/vnd.github+json\" \\\n -H \"Authorization: Bearer ${{ steps.dest_token_logic.outputs.token }}\" \\\n -H \"X-GitHub-Api-Version: 2022-11-28\" \\\n \"https://api.github.com/repos/${{ env.DEST_REPO_NAME }}/statuses/$SHA\" \\\n -d \"$PAYLOAD\" || true\n fi\n\n - name: Cleanup Temporary Files\n if: always()\n run: |\n set -euo pipefail\n rm -rf source-temp-* *.tar.gz include-list.txt include-effective.txt exclude-list.txt || true\n echo \"Cleanup done.\"\n" } ] } ] }, { "name": "backups", "type": "folder", "size": "1 bytes", "path": "backups", "children": [ { "name": "backup1", "type": "folder", "size": "1 bytes", "path": "backups/backup1", "children": [ { "name": "makefolder", "type": "file", "size": "1 bytes", "path": "backups/backup1/makefolder", "content": "\n" } ] } ] }, { "name": "c", "type": "folder", "size": "6,337 bytes", "path": "c", "children": [ { "name": "c++", "type": "folder", "size": "6,337 bytes", "path": "c/c++", "children": [ { "name": "1.cpp", "type": "file", "size": "6,337 bytes", "path": "c/c++/1.cpp", "content": "#include \n#include \n#include \n#include \n#define R return\n#define C esp_wifi_set_\n#define Z(v) if(v>3.14159)v-=6.28318;if(v<-3.14159)v+=6.28318\nstruct V{float x,p,a,o,lp,ld;int s,zc;};\nstruct L{uint32_t t;float p,v;};\nstruct F{float mn,mx;int pt;};\nV m[64];L h[200];int hx=0;uint32_t T;bool Ld=0,Cmp=0;\nWebServer S(80);int i;int8_t b[512];float l,d,cm,tm,iv,Q,al,av;F flt={0.5,50.0,5};\nuint8_t pkt[]={0xc0,0x00,0x00,0x00,0xff,0xff,0xff,0xff,0xff,0xff,0x00,0x00,0x00,0x00,0x00,0x00,0xff,0xff,0xff,0xff,0xff,0xff,0x00,0x00};\nconst char* P=R\"===(\n\n
\n

SENTINEL RADAR

\n
\n STATUS: INIT\n LIVE MASS: 0.0\n
\n
\n \n
\n
\n FILTERS (0 = OFF)\n
\n MIN SIZE:\n
\n
\n MAX SIZE:\n
\n
\n FAN/OSC:\n
\n \n
\n \n
\n)===\";\nvoid r(void*_,wifi_csi_info_t*o){if(o->len>256)R;memcpy(b,o->buf,o->len);l=0;cm=0;tm=0;\nfor(i=0;i<64;i++){float raw=atan2(b[i*2+1],b[i*2]);if(i){d=raw-l;Z(d);raw=m[i-1].x+d;}l=raw;\nfloat delta=raw-m[i].lp;if((delta>0&&m[i].ld<0)||(delta<0&&m[i].ld>0))m[i].zc++;\nif(m[i].zc>0&&rand()%10==0)m[i].zc--;m[i].lp=raw;m[i].ld=delta;\niv=raw-m[i].x;Q=iv*iv;if(Q>.5)Q=1.;else if(Q<.001)Q=1e-6;\nm[i].p+=Q;float k=m[i].p/(m[i].p+.1);m[i].x+=k*iv;m[i].p=(1.-k)*m[i].p;\nif(!Ld){m[i].a=m[i].x;m[i].o=m[i].x;m[i].s=0;}else{\nif(flt.pt>0&&m[i].zc>flt.pt){m[i].s=3;continue;}\nfloat rf=Cmp?m[i].o:m[i].a;float df=fabs(m[i].x-rf);\nif(df>.5){m[i].s=(m[i].p<.01)?1:2;if(m[i].s==2){cm+=i*df;tm+=df;}}\nelse{m[i].s=0;if(!Cmp)m[i].a=m[i].a*.999+m[i].x*.001;}}}\nav=0;if(Ld){if(flt.mn>0&&tm0&&tm>flt.mx)tm=0;\nif(tm>0){float p=cm/tm,v=0;int lx=hx?hx-1:199;al=p;av=1;\nif(h[lx].t>0){float dt=(millis()-h[lx].t)/1000.;if(dt>0)v=(p-h[lx].p)/dt;}\nh[hx]={millis(),p,v};hx=(hx+1)%200;}}\nvoid setup(){T=millis();WiFi.softAP(\"ESP32_SENTINEL\",\"\");\nfor(i=0;i<64;i++)m[i]={0,1,0,0,0,0,0,0};C csi_rx_cb(r,0);C csi(1);\nwifi_csi_config_t c={1,1,1,1,0,0,0};C csi_config(&c);S.on(\"/\",[](){S.send(200,\"text/html\",P);});\nS.on(\"/s\",[](){if(S.hasArg(\"m\"))Cmp=S.arg(\"m\").toInt();S.send(200,\"text/plain\",\"OK\");});\nS.on(\"/f\",[](){if(S.hasArg(\"n\"))flt.mn=S.arg(\"n\").toFloat();if(S.hasArg(\"x\"))flt.mx=S.arg(\"x\").toFloat();\nif(S.hasArg(\"p\"))flt.pt=S.arg(\"p\").toInt();S.send(200,\"text/plain\",\"OK\");});\nS.on(\"/d\",[](){String j=\"{\\\"s\\\":\\\"\"+String(Ld?\"ARMED\":\"LEARNING\")+\"\\\",\\\"sz\\\":\"+String(tm)+\",\\\"a\\\":\"+String(av)+\",\\\"l\\\":\"+String(al)+\",\\\"m\\\":[\";\nfor(i=0;i<64;i++)j+=(i?\",\":\"\")+String(\"{\\\"x\\\":\")+m[i].x+\",\\\"p\\\":\"+m[i].p+\",\\\"s\\\":\"+m[i].s+\"}\";j+=\"],\\\"h\\\":[\";\nint x=hx;for(int k=0;k<200;k++){if(h[x].t)j+=(k?\",\":\"\")+String(\"{\\\"t\\\":\")+h[x].t+\",\\\"p\\\":\"+h[x].p+\",\\\"v\\\":\"+h[x].v+\"}\";\nx=(x+1)%200;}S.send(200,\"application/json\",j+\"]}\");});S.begin();}\nvoid loop(){S.handleClient();if(!Ld&&millis()-T>60000){Ld=1;for(i=0;i<64;i++)m[i].o=m[i].x;}\nif(Ld){pkt[22]++;esp_wifi_80211_tx(WIFI_IF_AP,pkt,24,1);}delay(5);}\n" } ] } ] }, { "name": "containers", "type": "folder", "size": "255 bytes", "path": "containers", "children": [ { "name": "docker", "type": "folder", "size": "254 bytes", "path": "containers/docker", "children": [ { "name": "googlecloud", "type": "folder", "size": "254 bytes", "path": "containers/docker/googlecloud", "children": [ { "name": "cloudbuild.json", "type": "file", "size": "253 bytes", "path": "containers/docker/googlecloud/cloudbuild.json", "content": "{ \"steps\": [{\n \"name\": \"gcr.io/cloud-builders/docker\",\n \"args\": [\n \"build\",\n \"-f\",\n \"containers/docker/googlecloud/Dockerfile\",\n \"-t\",\n \"gcr.io/$PROJECT_ID/my-app\",\n \".\"]}]} //dont forget to set trigger in free time \n" }, { "name": "Dockerfile", "type": "file", "size": "1 bytes", "path": "containers/docker/googlecloud/Dockerfile", "content": "\n" } ] } ] }, { "name": "vm", "type": "folder", "size": "1 bytes", "path": "containers/vm", "children": [ { "name": "makemeafolder.txt", "type": "file", "size": "1 bytes", "path": "containers/vm/makemeafolder.txt", "content": "\n" } ] } ] }, { "name": "google_apps_script", "type": "folder", "size": "7,840 bytes", "path": "google_apps_script", "children": [ { "name": "1.gs", "type": "file", "size": "6,729 bytes", "path": "google_apps_script/1.gs", "content": "function auto_share_for_connected_ai_studio() {\n var force_full_check = 0;\n var stop_at_error = 0;\n var alert_at_error = 1;\n var mail_daily_report = 1;\n var mail_monthly_report = 1;\n var mail_target = 0;\n var share_notification_silencer = 0;\n\n var lock = LockService.getScriptLock();\n if (!lock.tryLock(30000)) return;\n\n try {\n var p = PropertiesService.getScriptProperties();\n var team = [\"mail12@gmail.com\", \"mail13@gmail.com\", \"mail14@gmail.com\", \"mail15@gmail.com\", \"mail16@gmail.com\"];\n \n var me = Session.getActiveUser().getEmail();\n var work_team = [];\n for(var i=0; i 5) {\n send_mail_logic(mail_target, team, \"daily report at \" + last_log_day + \" for auto_share_for_connected_ai_studio\", \"Activity Log:\\n\\n\" + y_content);\n }\n }\n p.setProperty(\"LOG_DATE_DAY\", today_str);\n } else if (!last_log_day) p.setProperty(\"LOG_DATE_DAY\", today_str);\n\n var last_log_month = p.getProperty(\"LOG_DATE_MONTH\");\n if (last_log_month && last_log_month !== month_str) {\n if (mail_monthly_report === 1) {\n var m_content = read_drive_file(storage_id, \"monthly_log_\" + last_log_month + \".txt\");\n if (m_content && m_content.length > 5) {\n var target = (mail_target === 0) ? 1 : mail_target;\n send_mail_logic(target, team, \"monthly report at \" + last_log_month + \" for auto_share_for_connected_ai_studio\", \"Monthly Summary:\\n\\n\" + m_content);\n }\n }\n p.setProperty(\"LOG_DATE_MONTH\", month_str);\n } else if (!last_log_month) p.setProperty(\"LOG_DATE_MONTH\", month_str);\n\n var k = \"T_V8\";\n var qt = p.getProperty(k) || \"1970-01-01T00:00:00.000Z\";\n if (force_full_check === 1) qt = \"1970-01-01T00:00:00.000Z\";\n var st = qt, total_up = 0;\n var f = DriveApp.getFoldersByName(\"Google AI Studio\");\n var silent = (share_notification_silencer === 2 || (share_notification_silencer === 1 && force_full_check === 1));\n var run_errors = [];\n\n while (f.hasNext()) {\n var fid = f.next().getId(), page;\n do {\n var res = Drive.Files.list({\n q: \"'\" + fid + \"' in parents and createdTime > '\" + qt + \"' and trashed = false\",\n orderBy: \"createdTime\",\n fields: \"nextPageToken, files(id, name, createdTime, permissions(emailAddress))\",\n pageToken: page\n });\n var list = res.files;\n if (!list || list.length === 0) break;\n\n for (var i = 0; i < list.length; i++) {\n var item = list[i], cur = [];\n if (item.permissions) for (var j = 0; j < item.permissions.length; j++) cur.push(item.permissions[j].emailAddress);\n var add = [];\n for (var x = 0; x < work_team.length; x++) if (cur.indexOf(work_team[x]) < 0) add.push(work_team[x]);\n\n if (add.length > 0) {\n try {\n if (silent) {\n for (var z = 0; z < add.length; z++) {\n Drive.Permissions.create({ role: 'editor', type: 'user', emailAddress: add[z] }, item.id, { sendNotificationEmail: false });\n }\n } else {\n DriveApp.getFileById(item.id).addEditors(add);\n }\n total_up++;\n } catch (e) {\n var err = \"File: \" + item.name + \" (\" + item.id + \") Error: \" + e.message;\n console.error(err);\n run_errors.push(err);\n if (stop_at_error === 1) throw new Error(\"STOPPED by setting: \" + err);\n }\n }\n if (item.createdTime > st) st = item.createdTime;\n }\n page = res.nextPageToken;\n } while (page);\n }\n\n if (st !== qt) p.setProperty(k, st);\n \n if (total_up > 0) {\n var ts = new Date().toTimeString().slice(0, 5);\n var entry = \"\\n[\" + ts + \"] Shared \" + total_up + \" files.\";\n append_to_drive_file(storage_id, \"daily_log_\" + today_str + \".txt\", entry);\n append_to_drive_file(storage_id, \"monthly_log_\" + month_str + \".txt\", \"\\n[\" + today_str + \" \" + ts + \"] Shared \" + total_up + \" files.\");\n }\n \n if (run_errors.length > 0 && alert_at_error === 1) {\n trigger_calendar_popup(\"🚨 Auto-Share Error!\", \"Errors found during run:\\n\" + run_errors.join(\"\\n\"));\n }\n\n } catch (critical_e) {\n if (alert_at_error === 1) trigger_calendar_popup(\"🚨 Script CRASHED!\", critical_e.message);\n } finally {\n lock.releaseLock();\n }\n}\n\nfunction trigger_calendar_popup(title, description) {\n try {\n var now = new Date();\n var end = new Date(now.getTime() + 60000); \n var event = CalendarApp.getDefaultCalendar().createEvent(title, now, end, { description: description });\n event.addPopupReminder(0); \n } catch(e) {\n console.log(\"Calendar Alert Failed: \" + e.message);\n }\n}\n\nfunction append_to_drive_file(folderId, filename, text) {\n try {\n var folder = DriveApp.getFolderById(folderId);\n var files = folder.getFilesByName(filename);\n if (files.hasNext()) {\n var file = files.next();\n file.setContent(file.getBlob().getDataAsString() + text);\n } else {\n folder.createFile(filename, text);\n }\n } catch(e) {}\n}\n\nfunction read_drive_file(folderId, filename) {\n try {\n var folder = DriveApp.getFolderById(folderId);\n var files = folder.getFilesByName(filename);\n if (files.hasNext()) return files.next().getBlob().getDataAsString();\n } catch(e) {}\n return \"\";\n}\n\nfunction send_mail_logic(mode, team_list, subject, body) {\n var rec = \"\";\n if (mode === 0) rec = Session.getActiveUser().getEmail();\n else if (mode === 1) rec = team_list.join(\",\");\n else if (typeof mode === \"string\" && mode.includes(\"@\")) rec = mode;\n else rec = Session.getActiveUser().getEmail();\n \n if (rec) MailApp.sendEmail(rec, subject, body);\n}\n" }, { "name": "2.gs", "type": "file", "size": "1,111 bytes", "path": "google_apps_script/2.gs", "content": "function auto_share_for_connected_ai_studio(){\nvar force_full_check=0;\nvar p=PropertiesService.getScriptProperties();\nvar k=\"T_V8\";\nvar r=p.getProperty(k);\nvar qt=\"1970-01-01T00:00:00.000Z\";\nif(r&&force_full_check===0)qt=r;\nvar st=qt;\nvar f=DriveApp.getFoldersByName(\"Google AI Studio\");\nvar m=[\"2@gmail.com\",\"3@gmail.com\",\"4@gmail.com\",\"5@gmail.com\",\"6@gmail.com\"];\nwhile(f.hasNext()){\nvar fid=f.next().getId(),page;\ndo{\nvar res=Drive.Files.list({q:\"'\"+fid+\"' in parents and createdTime > '\"+qt+\"' and trashed = false\",orderBy:\"createdTime\",fields:\"nextPageToken, files(id, createdTime, permissions(emailAddress))\",pageToken:page});\nvar list=res.files;\nif(!list||list.length===0)break;\nfor(var i=0;i0)DriveApp.getFileById(item.id).addEditors(add);\nif(item.createdTime>st)st=item.createdTime;\n}\npage=res.nextPageToken;\n}while(page);\n}\nif(st!==qt)p.setProperty(k,st);\n}\n" } ] }, { "name": "kotlin", "type": "folder", "size": "13,943 bytes", "path": "kotlin", "children": [ { "name": "old", "type": "folder", "size": "5,658 bytes", "path": "kotlin/old", "children": [ { "name": "1old1.kt", "type": "file", "size": "5,658 bytes", "path": "kotlin/old/1old1.kt", "content": "REDACTED" } ] }, { "name": "1.kt", "type": "file", "size": "8,285 bytes", "path": "kotlin/1.kt", "content": "package a.htmlapprealizer;import android.app.*;import android.os.*;import android.view.*;import android.webkit.*;import android.widget.*;import android.text.*;import android.text.style.*;import android.graphics.*;import org.json.*;import java.util.*;import java.lang.reflect.*;class master:Activity(){val H=ArrayList();val L=Array(3){HashSet()};val D=HashSet();lateinit var w:WebView;lateinit var b:Button;var m=0;var s=true;var A=false;var NET=false;var PW=\"\";val P by lazy{getSharedPreferences(\"Z\",0)};fun S(k:String,v:Any)=with(P.edit()){if(v is Boolean)putBoolean(k,v)else putString(k,v.toString());apply()};override fun onCreate(k:Bundle?){super.onCreate(k);H.add(this);(0..2).forEach{i->L[i].addAll(P.getString(\"$i\",\"\")!!.split(\",\"))};m=P.getInt(\"M\",0);s=P.getBoolean(\"S\",true);NET=P.getBoolean(\"N\",false);PW=P.getString(\"W\",\"\")!!;D.addAll(P.getString(\"D\",\"\")!!.split(\",\"));val f=FrameLayout(this);b=Button(this).apply{text=\"⚙\";setTextColor(-1);setBackgroundColor(0x99000000.toInt());layoutParams=FrameLayout.LayoutParams(120,120,53);setOnClickListener{E()}};w=WebView(this).apply{settings.javaScriptEnabled=true;settings.domStorageEnabled=true;addJavascriptInterface(this@master,\"K\");webViewClient=object:WebViewClient(){override fun onPageStarted(v:WebView?,u:String?,b:Bitmap?){A=false};override fun shouldInterceptRequest(v:WebView?,r:WebResourceRequest?):WebResourceResponse?{val h=r?.url?.host;return if(NET&&r?.url?.scheme !in listOf(\"file\",\"data\")&&!(h!=null&&D.any{it!=\"\"&&(h==it||h.endsWith(\".$it\"))}))WebResourceResponse(null,null,null)else null}}};f.addView(w);f.addView(b);setContentView(f);v(m);w.loadDataWithBaseURL(null,\"\"\"HTML/App Realizer
\"\"\",\"text/html\",\"utf-8\",null)};override fun onResume(){super.onResume();if(m==1)v(0)};override fun onDestroy(){super.onDestroy();w.destroy();H.clear()};override fun onKeyDown(k:Int,e:KeyEvent?):Boolean{if(k==24){b.visibility=0;Handler(Looper.getMainLooper()).postDelayed({v(m)},3000);return false};return super.onKeyDown(k,e)};fun v(n:Int){m=n;if(m==3)S(\"M\",3);b.visibility=if(m==0)0 else 8};fun ok()=!s&&(PW==\"\"||A);fun E(){val sv=ScrollView(this);val l=LinearLayout(this).apply{orientation=1;setPadding(50,50,50,50);setBackgroundColor(-13421773)};sv.addView(l);fun T(e:EditText,f:(String)->Unit)=e.addTextChangedListener(object:TextWatcher{override fun afterTextChanged(s:Editable?)=f(s.toString());override fun beforeTextChanged(s:CharSequence?,x:Int,y:Int,z:Int){};override fun onTextChanged(s:CharSequence?,x:Int,y:Int,z:Int){}});fun U(t:String,c:Int,h:String,v:String,a:(Button)->Unit,x:(String)->Unit){val r=LinearLayout(this);val B=Button(this).apply{text=t;setTextColor(c);setBackgroundColor(0);setOnClickListener{a(this)}};val E=EditText(this).apply{hint=h;setTextColor(-1);setText(v);layoutParams=LinearLayout.LayoutParams(0,-2,1f)};T(E,x);r.addView(B);r.addView(E);l.addView(r)};U(\"Sandbox: ${if(s)\"ON\" else \"OFF\"}\",if(s)-16711936 else -65536,\"Password\",PW,{s=!s;S(\"S\",s);it.text=\"Sandbox: ${if(s)\"ON\" else \"OFF\"}\";it.setTextColor(if(s)-16711936 else -65536)}){PW=it;S(\"W\",it);A=false};val tv=TextView(this).apply{setTextColor(-1);textSize=16f;movementMethod=android.text.method.LinkMovementMethod.getInstance();text=SpannableStringBuilder().apply{val m=mapOf(\"unhide\" to 0,\"focus\" to 1,\"open\" to 2,\"perma\" to 3);\"unhide/hide settings:till next focus/open/perma\".split(Regex(\"(?<=[a-z])(?=[/:])|(?<=[/:])(?=[a-z])\")).forEach{w->if(m.containsKey(w)){val s=length;append(\" $w \");setSpan(object:ClickableSpan(){override fun onClick(v:View){v(m[w]!!)};override fun updateDrawState(d:TextPaint){d.color=-1;d.bgColor=-12303292;d.isUnderlineText=false}},s,length,33)}else append(w)}}};l.addView(tv);U(\"Cut Internet: ${if(NET)\"ON\" else \"OFF\"}\",if(NET)-65536 else -16711936,\"Domain Exceptions\",D.joinToString(\",\"),{NET=!NET;S(\"N\",NET);it.text=\"Cut Internet: ${if(NET)\"ON\" else \"OFF\"}\";it.setTextColor(if(NET)-65536 else -16711936)}){D.clear();D.addAll(it.split(\",\"));S(\"D\",it)};val n=arrayOf(\"Whitelist\",\"Graylist\",\"Blacklist\");(0..2).forEach{i->l.addView(TextView(this).apply{text=n[i];setTextColor(-1)});val e=EditText(this).apply{setText(L[i].joinToString(\",\"));setTextColor(-1);setBackgroundColor(-12303292)};T(e){L[i].clear();L[i].addAll(it.split(\",\"));S(\"$i\",it)};l.addView(e)};AlertDialog.Builder(this).setView(sv).show()};fun K(i:Int,s:String)=L[i].contains(\"ALL\")||L[i].any{it!=\"\"&&s.contains(it)};fun C(n:String)=try{Class.forName(n)}catch(e:Exception){null};fun R(t:Any?,n:String,j:String,y:Int,cI:Int):String{if(!ok())return \"E:SEC\";try{val c=if(t is Class<*>)t else t?.javaClass?:return \"E:NUL\";val a=JSONArray(j);val sig=\"$n ${c.name}\".lowercase();if(y<4&&K(0,sig))return \"E:BL\";val g=if(y>3||K(1,sig))null else if(K(2,sig))\"ALL\" else L[2].find{it!=\"\"&&sig.contains(it)}?:if(L[0].isEmpty()&&L[1].isEmpty())\"ALL\" else null;if(g!=null){val i=(System.currentTimeMillis()%999).toInt();runOnUiThread{AlertDialog.Builder(this).setTitle(\"REQ:$sig\").setPositiveButton(\"1\"){_,_->cb(cI,Ex(t,n,a,y))}.setNeutralButton(\"OK\"){_,_->L[1].add(g);S(\"1\",L[1].joinToString(\",\"));cb(cI,Ex(t,n,a,y))}.setNegativeButton(\"NO\"){_,_->L[0].add(g);S(\"0\",L[0].joinToString(\",\"));cb(cI,\"E:BL\")}.show()};return \"W:$i\"};return Ex(t,n,a,y)}catch(e:Exception){return \"E:$e\"}};fun Ex(t:Any?,n:String,a:JSONArray,y:Int):String{try{val c=if(t is Class<*>)t else t!!.javaClass;val l=a.length();if(y==2)return ret(c.getField(n).get(t));if(y==3){val f=c.getField(n);f.set(t,cv(a.get(0),f.type));return \"OK\"};if(y==4)return ret(java.lang.reflect.Array.newInstance(c,a.getInt(0)));if(y==5)return ret(Proxy.newProxyInstance(c.classLoader,arrayOf(c)){_,m,r->val g=r?.map{ret(it)}?.joinToString(\",\")?:\"\";runOnUiThread{w.evaluateJavascript(\"window.onL($l,'${m.name}','[$g]')\",null)};null});val ms=if(y==1)c.constructors else c.methods;for(m in ms){val pts=m.parameterTypes;if((y==0&&m.name!=n)||pts.size!=l)continue;try{val v=Array(l){i->cv(a.get(i),pts[i])};val r=if(y==1)(m as Constructor<*>).newInstance(*v)else(m as Method).invoke(if(t is Class<*>)null else t,*v);return ret(r)}catch(e:Exception){continue}};return \"E:SIG\"}catch(e:Exception){return if(e is InvocationTargetException)\"E:${e.targetException}\"else\"E:$e\"}};fun cv(o:Any,t:Class<*>)=if(o is String&&o.startsWith(\"P\"))H.getOrNull(o.substring(1).toIntOrNull()?:-1)else if(o is Number)when(t){Int::class.javaPrimitiveType->o.toInt();Long::class.javaPrimitiveType->o.toLong();Float::class.javaPrimitiveType->o.toFloat();Double::class.javaPrimitiveType->o.toDouble();else->o}else if(t==Boolean::class.javaPrimitiveType)o.toString().toBoolean()else o;fun ret(o:Any?)=if(o==null||o is Unit)\"V\" else if(o is Number||o is String||o is Boolean)\"V$o\" else{H.add(o);\"P${H.size-1}\"};fun cb(i:Int,r:String)=if(i>0)w.evaluateJavascript(\"window.onC($i,'${r.replace(\"\\\\\",\"\\\\\\\\\").replace(\"'\",\"\\\\'\").replace(\"\\n\",\"\\\\n\")}')\",null)else Unit;@JavascriptInterface fun sz()=H.size;@JavascriptInterface fun cls(){H.clear();H.add(this)};@JavascriptInterface fun c(n:String)=try{if(ok()){H.add(Class.forName(n));\"P${H.size-1}\"}else\"E:SEC\"}catch(e:Exception){\"E:$e\"};@JavascriptInterface fun n(c:String,j:String)=R(C(c),\"\",j,1,0);@JavascriptInterface fun x(p:Int,n:String,j:String)=R(H.getOrNull(p),n,j,0,0);@JavascriptInterface fun u(p:Int,n:String,j:String,cb:Int){runOnUiThread{cb(cb,R(H.getOrNull(p),n,j,0,cb))}};@JavascriptInterface fun g(p:Int,f:String)=R(H.getOrNull(p),f,\"[]\",2,0);@JavascriptInterface fun s(p:Int,f:String,v:String)=R(H.getOrNull(p),f,\"[$v]\",3,0);@JavascriptInterface fun a(t:String,l:Int)=R(C(t),\"\", \"[$l]\", 4, 0);@JavascriptInterface fun p(t:String,id:Int)=R(C(t),\"\", \"[$id]\", 5, 0);@JavascriptInterface fun del(p:Int){if(ok()&&p>0)H[p]=null};@JavascriptInterface fun auth(p:String):Boolean{A=(p==PW);return A}}\n" } ] }, { "name": "old", "type": "folder", "size": "183,838 bytes", "path": "old", "children": [ { "name": "hybrid_filesold0.html", "type": "file", "size": "1,437 bytes", "path": "old/hybrid_filesold0.html", "content": "REDACTED" }, { "name": "indexold1.html", "type": "file", "size": "1,437 bytes", "path": "old/indexold1.html", "content": "REDACTED" }, { "name": "repo_to_jsonold1.html", "type": "file", "size": "9,621 bytes", "path": "old/repo_to_jsonold1.html", "content": "REDACTED" }, { "name": "repo_to_jsonold2.html", "type": "file", "size": "18,827 bytes", "path": "old/repo_to_jsonold2.html", "content": "REDACTED" }, { "name": "repo_to_jsonold3.html", "type": "file", "size": "14,362 bytes", "path": "old/repo_to_jsonold3.html", "content": "REDACTED" }, { "name": "website_changerold1.html", "type": "file", "size": "12,028 bytes", "path": "old/website_changerold1.html", "content": "REDACTED" }, { "name": "website_changerold2.html", "type": "file", "size": "16,756 bytes", "path": "old/website_changerold2.html", "content": "REDACTED" }, { "name": "website_changerold3.html", "type": "file", "size": "24,069 bytes", "path": "old/website_changerold3.html", "content": "REDACTED" }, { "name": "website_changerold4.html", "type": "file", "size": "24,753 bytes", "path": "old/website_changerold4.html", "content": "REDACTED" }, { "name": "website_changerold5.html", "type": "file", "size": "28,317 bytes", "path": "old/website_changerold5.html", "content": "REDACTED" }, { "name": "website_changerold6.html", "type": "file", "size": "32,231 bytes", "path": "old/website_changerold6.html", "content": "REDACTED" } ] }, { "name": "python", "type": "folder", "size": "5,092 bytes", "path": "python", "children": [ { "name": "clean_c_executer_2.py", "type": "file", "size": "2,032 bytes", "path": "python/clean_c_executer_2.py", "content": "import os,re,sys,subprocess,tempfile\ndef compile_and_run(c_code,source_name):\n try:\n print(f\"-> Compiling C code from '{source_name}'...\")\n with tempfile.TemporaryDirectory() as d:\n c,e=os.path.join(d,\"c.c\"),os.path.join(d,\"e.out\" if os.name!='nt' else \"e.exe\")\n with open(c,'w')as f:f.write(c_code)\n cp=subprocess.run(['gcc','-O3',c,'-o',e],capture_output=True,text=True)\n if cp.returncode!=0:\n print(\"\\n\"+\"=\"*60);print(\"--- C COMPILATION FAILED ---\");print(f\"Source: {source_name}\\n\\n[GCC Error]\\n{cp.stderr.strip()}\\n\\n[Code]\")\n for i,line in enumerate(c_code.strip().splitlines(),1):print(f\"{i: >3}: {line}\")\n print(\"=\"*60);sys.exit(1)\n print(\"-> Compilation successful. Running program...\")\n rp=subprocess.run([e],cwd=d,capture_output=True,text=True)\n if rp.returncode!=0:print(f\"\\n[!] C Program Crashed (Code {rp.returncode}). Memory dumps trapped and destroyed.\")\n print(\"\\n--- Program Output ---\\n\"+rp.stdout.strip()+\"\\n----------------------\")\n finally:\n print(\"\\n-> Memory sandbox unlinked. Zero artifacts remain.\")\ndef main():\n print(\"--- C Code Runner Initialized (Fallback Order: Auto > File > String) ---\")\n args=sys.argv;num_cli_args=len(args)-1; c_code, source_name = None, None\n if num_cli_args == 0 or num_cli_args > 2:\n files={int(m.group(1)):f for f in os.listdir('.') if(m:=re.match(r'input_file_(\\d+)\\.py',f))and int(m.group(1))!=0}\n if files:\n source_name=files[max(files.keys())];print(f\"Mode: Auto-Discovery ('{source_name}')\")\n with open(source_name, 'r') as f: c_code = f.read()\n elif num_cli_args == 1 and args[1] != '0' and os.path.exists(args[1]):\n source_name=args[1];print(f\"Mode: Direct File ('{source_name}')\")\n with open(source_name, 'r') as f: c_code = f.read()\n elif num_cli_args == 2 and args[1] == '0':\n print(\"Mode: Direct String Injection\"); c_code, source_name = args[2], \"direct_string\"\n if c_code is not None: compile_and_run(c_code, source_name)\n else: print(\"Error: No valid input found.\",file=sys.stderr);sys.exit(1)\nif __name__==\"__main__\":main()\n" }, { "name": "clean_c_executer.py", "type": "file", "size": "3,060 bytes", "path": "python/clean_c_executer.py", "content": "try:\n with open(\"input_file_0.py\", \"w\") as f:\n f.write(r\"\"\"import os,re,sys,subprocess\ndef compile_and_run(c_code,source_name):\n executable_name=f\"delete_{os.path.splitext(os.path.basename(source_name))[0]}\";temp_c_file=\"_temp_source.c\"\n try:\n print(f\"-> Compiling C code from '{source_name}'...\")\n with open(temp_c_file,'w')as f:f.write(c_code)\n cp=subprocess.run(['gcc',temp_c_file,'-o',executable_name],capture_output=True,text=True)\n if cp.returncode!=0:\n print(\"\\n\"+\"=\"*60);print(\"--- C COMPILATION FAILED ---\");print(f\"Source: {source_name}\\n\\n[GCC Compiler Error Message]\");print(cp.stderr.strip());print(\"\\n[Problematic Code with Line Numbers]\")\n for i,line in enumerate(c_code.strip().splitlines(),1):print(f\"{i: >3}: {line}\")\n print(\"=\"*60);sys.exit(1)\n print(\"-> Compilation successful. Running program...\")\n rp=subprocess.run([f'./{executable_name}'],check=True,capture_output=True,text=True)\n print(\"\\n--- Program Output ---\\n\"+rp.stdout.strip()+\"\\n----------------------\")\n finally:\n print(\"\\n-> Cleaning up artifacts...\")\n cl=False\n if os.path.exists(temp_c_file):os.remove(temp_c_file);cl=True\n if os.path.exists(executable_name):os.remove(executable_name);cl=True\n if cl:print(\" Cleanup complete.\")\ndef main():\n print(\"--- C Code Runner Initialized (Fallback Order: Auto > File > String) ---\")\n args=sys.argv;num_cli_args=len(args)-1\n c_code, source_name = None, None\n if num_cli_args == 0 or num_cli_args > 2:\n files={int(m.group(1)):f for f in os.listdir('.') if(m:=re.match(r'input_file_(\\d+)\\.py',f))and int(m.group(1))!=0}\n if files:\n print(\"Mode: Automatic Discovery\")\n source_name=files[max(files.keys())];print(f\"Found target: '{source_name}'\")\n with open(source_name, 'r') as f: c_code = f.read()\n elif num_cli_args == 1 and args[1] != '0' and os.path.exists(args[1]):\n source_name=args[1];print(f\"Mode: Direct File Target ('{source_name}')\")\n with open(source_name, 'r') as f: c_code = f.read()\n elif num_cli_args == 2 and args[1] == '0':\n print(\"Mode: Direct String Injection\")\n c_code, source_name = args[2], \"direct_string\"\n if c_code is not None:\n compile_and_run(c_code, source_name)\n else:\n print(\"\\nError: No action could be taken.\",file=sys.stderr)\n print(\" - Auto-discovery failed: No 'input_file_X.py' files found.\",file=sys.stderr)\n print(\" - Argument Error: Check arguments for File or String mode.\",file=sys.stderr);sys.exit(1)\nif __name__==\"__main__\":main()\n\"\"\")\n with open(\"input_file_100.py\", \"w\") as f:\n f.write('#include \\nint main(){printf(\"Result is: %d\\\\n\",2+2);return 0;}')\n print(\"Successfully created 'input_file_0.py' and 'input_file_100.py'.\")\n print(\"\\nTo run the code, use one of the following commands:\")\n print(\"1. (Auto-Search): python3 input_file_0.py\")\n print(\"2. (Direct File): python3 input_file_0.py input_file_100.py\")\n print(\"3. (Direct String): python3 input_file_0.py 0 '#include \\\\nint main(){printf(\\\\\\\"Hello!\\\\n\\\\\\\");}'\")\nexcept IOError as e:\n print(f\"An error occurred during file creation: {e}\")\n" } ] }, { "name": "scratchpad", "type": "folder", "size": "258,609 bytes", "path": "scratchpad", "children": [ { "name": "game", "type": "folder", "size": "43,062 bytes", "path": "scratchpad/game", "children": [ { "name": "helper", "type": "folder", "size": "43,062 bytes", "path": "scratchpad/game/helper", "children": [ { "name": "dimensional_vector_optimizer.html", "type": "file", "size": "37,994 bytes", "path": "scratchpad/game/helper/dimensional_vector_optimizer.html", "content": "\n\n\n \n \n Dimensional Vector Extraction Optimizer\n \n\n\n\n

Dimensional Vector Extraction Optimizer

\n \n
\n \n \n \n \n \n \n
\n\n
\n
dimension count of hyperspace:
\n \n
\n \n \n main vector in hyperspace↓ \n
\n [\n \n ]\n
\n
\n
\n
\n\n
\n
\n \n \n vectors to be extracted:\n
\n
\n
\n\n
\n \n
\n\n \n
\n
\n faster for smaller size: Basic DFS\n \n faster for bigger size: Surrogate ILP\n
\n\n
\n \n \n
\n\n
\n
Status: Calculating deeply...
\n
Nodes Explored: 0
\n
Best Score Found: 0
\n \n
\n\n
\n
\n\n\n\n\n" }, { "name": "healoptimal.html", "type": "file", "size": "5,068 bytes", "path": "scratchpad/game/helper/healoptimal.html", "content": "Logistics Calc\n\n\n

🚢 Config

\n
\n

🎯 Target Finder

\n

📋 Lists



\n
\n\n" } ] } ] }, { "name": "hatch", "type": "folder", "size": "110,651 bytes", "path": "scratchpad/hatch", "children": [ { "name": "primitive_dimensional_subspace.html", "type": "file", "size": "29,981 bytes", "path": "scratchpad/hatch/primitive_dimensional_subspace.html", "content": "\n\n\n \n \n Strict Step-by-Step Subspace Visualizer\n \n\n\n\n

Strict Step-by-Step Subspace Visualizer

\n

Nothing skipped. Watch every projection, pivot, and extraction frame by frame.

\n\n
\n \n
\n

Configuration

\n
\n \n \n
\n
\n \n \n
\n
\n \n \n
\n \n
\n
\n \n
\n \n x\n
\n
\n \n Scale: 0.1x (Slow) to 1000x (Instant)\n
\n\n
\n \n \n
\n\n
\n
Complexity: O(1)
\n
Math Operations: 0
\n
\n\n

Terminal Log

\n
\n
\n\n \n
\n \n
Waiting to initialize...
\n\n
\n
\n Working Plane (Known Vectors)\n \n
\n
\n
\n\n
\n
\n Extracted Orthogonal Subspace (Null Space)\n \n
\n
\n
\n\n
\n

Mathematical Proof (Dot Product Check)

\n
\n
\n
\n
\n\n \n\n\n" }, { "name": "primitive_risky_website_inspector.html", "type": "file", "size": "13,119 bytes", "path": "scratchpad/hatch/primitive_risky_website_inspector.html", "content": "\n\n\n \n \n Cyber Forensics Sandbox v5.0 (Optimized)\n \n\n\n\n

🛡️ Cyber Forensics Sandbox v5.0 (Optimized Core)

\n \n
\n
\n

📡 Option 1: Live Fetch URL

\n
Fetches code via CORS proxies. Best for generating a 1:1 Visual Sandbox layout.
\n \n \n
\n\n
\n

📂 Option 2: Upload Offline File (.mhtml)

\n
Extracts code from mobile downloads. Best for bypassing IP blocks and cloakers.
\n \n \n
\n
\n\n
Ready. Select an option above.
\n\n
\n \n \n \n
\n\n
\n \n \n \n
\n\n \n\n\n" }, { "name": "qrreaderold0.html", "type": "file", "size": "62,047 bytes", "path": "scratchpad/hatch/qrreaderold0.html", "content": "REDACTED" }, { "name": "sw.js", "type": "file", "size": "5,504 bytes", "path": "scratchpad/hatch/sw.js", "content": "self.oninstall = event => event.waitUntil(self.skipWaiting());\nself.onactivate = event => event.waitUntil(self.clients.claim());\n\nconst S = self.evalsandboxperclientv1 = { \n errors: Object.assign(new Array(10001).fill(null), { 0: 0 }),\n routeMap: {},\n status: { locked: false }, \n ...self.evalsandboxperclientv1 \n};\n\nconst trap = (fn, ev) => { \n try { \n let p = fn(ev); \n if (p?.catch) p.catch(x => S.errors[S.errors[0]++ % 10000 + 1] = x.stack || String(x)); \n return p; \n } catch (x) { \n S.errors[S.errors[0]++ % 10000 + 1] = x.stack || String(x); \n } \n};\n\nconst manifest = event => {\n if (S[event.type]) {\n let prom = trap(S[event.type], event);\n if (prom) event.waitUntil(prom);\n }\n};\nself.onpush = self.onsync = self.onnotificationclick = self.onnotificationclose = manifest;\n\nself.onmessage = async event => {\n let ES = event.data?.evalsandboxperclientv1;\n if (!ES) return;\n\n let cid = event.source.id;\n let curl = new URL(event.source.url).pathname;\n\n let follow = ES.status?.multilinkfollow ?? S[cid]?.status?.multilinkfollow ?? S[curl]?.status?.multilinkfollow ?? 0;\n let PTR = follow ? cid : curl;\n\n let C = S[PTR] ??= {};\n let ESS = C.status ??= {};\n\n ESS.acknowledgement = ES.status?.acknowledgement ?? ESS.acknowledgement ?? null;\n ESS.persistence = ES.status?.persistence ?? ESS.persistence ?? 0;\n ESS.multilinkfollow = ES.status?.multilinkfollow ?? ESS.multilinkfollow ?? 0;\n ESS.wakelock = ES.status?.wakelock ?? ESS.wakelock ?? 0;\n ESS.revivingpayload = ES.status?.revivingpayload ?? ESS.revivingpayload ?? 0;\n ESS.haveidied = ES.status?.haveidied ?? ESS.haveidied ?? 1;\n ESS.clientid = ES.status?.clientid ?? ESS.clientid ?? cid;\n ESS.clienturl = ES.status?.clienturl ?? ESS.clienturl ?? curl;\n\n for (let key in ES) {\n if (key !== 'status') C[key] = ES[key] ?? C[key];\n }\n\n ES.status = ESS; \n S.routeMap[ESS.clientid] = ESS.clienturl;\n\n if (ES.payload !== undefined) {\n C.payload = ES.payload;\n ESS.haveidied = 1;\n } else if (ESS.revivingpayload && !C.fn && ESS.persistence) {\n let diskRecord = await (await caches.open('S')).match(ESS.clienturl);\n if (diskRecord) {\n C.payload = await diskRecord.text();\n ESS.haveidied = 1;\n }\n }\n\n if (C.payload && (!C.fn || ES.payload !== undefined)) {\n C.fn = new (async()=>{}).constructor('event', C.payload);\n trap(C.fn, event);\n ESS.haveidied = 0;\n\n if (ESS.persistence && ES.payload !== undefined) {\n await (await caches.open('S')).put(ESS.clienturl, new Response(C.payload));\n }\n ESS.acknowledgement = S.errors[0] === ES.errors?.[0]; // Ack is true if errors didn't increase\n if (!ESS.acknowledgement) ES.last_error = S.errors[(S.errors[0] === 0 ? 1 : S.errors[0]) % 10000 || 10000];\n }\n\n if (ES.history) ES.errors = S.errors;\n event.source.postMessage({ evalsandboxperclientv1: ES });\n\n if (ESS.wakelock && !S.status.locked) {\n S.status.locked = true;\n event.waitUntil((async () => {\n while (await self.clients.matchAll().then(c => c.length > 0)) {\n await new Promise(r => setTimeout(r, 10000));\n }\n S.routeMap = {}; \n S.status.locked = false;\n })());\n }\n};\n\nself.onfetch = event => {\n let routeFallback = new URL(event.request.mode === 'navigate' ? event.request.url : (event.request.referrer || event.request.url)).pathname;\n let PTR = S[event.clientId] ? event.clientId : (S.routeMap[event.clientId] || routeFallback);\n\n if (S[PTR] && S[PTR].fn) {\n trap(S[PTR].fn, event);\n } else {\n event.respondWith((async () => {\n let trueRoute = PTR;\n if (!S.routeMap[event.clientId] && event.clientId) {\n let activeClient = await self.clients.get(event.clientId);\n if (activeClient) {\n trueRoute = new URL(activeClient.url).pathname;\n S.routeMap[event.clientId] = trueRoute;\n }\n }\n \n let diskRecord = await (await caches.open('S')).match(trueRoute);\n \n if (diskRecord) {\n let rawCode = await diskRecord.text();\n \n S[trueRoute] ??= {};\n S[trueRoute].status ??= { persistence: 1, multilinkfollow: 0, clienturl: trueRoute };\n S[trueRoute].status.haveidied = 1;\n \n S[trueRoute].payload = rawCode;\n S[trueRoute].fn = new (async()=>{}).constructor('event', rawCode);\n \n let capturedPromise;\n let proxyEvent = new Proxy(event, {\n get: (target, prop) => {\n if (prop === 'respondWith') return (prom) => { capturedPromise = prom; };\n return typeof target[prop] === 'function' ? target[prop].bind(target) : target[prop];\n }\n });\n \n trap(S[trueRoute].fn, proxyEvent);\n S[trueRoute].status.haveidied = 0;\n \n if (capturedPromise) return await capturedPromise;\n }\n \n return fetch(event.request);\n })().catch(() => fetch(event.request)));\n }\n};\n" } ] }, { "name": "modeltester", "type": "folder", "size": "2,539 bytes", "path": "scratchpad/modeltester", "children": [ { "name": "prompts", "type": "folder", "size": "2,539 bytes", "path": "scratchpad/modeltester/prompts", "children": [ { "name": "log1.txt", "type": "file", "size": "2,539 bytes", "path": "scratchpad/modeltester/prompts/log1.txt", "content": "hi, check this js code, be hyper aware it's tricky, here are main points I want(compulsory)but you can self add more if you wanna mention(preferred)\n\nnote: you can use unlimited compute and unlimited response length(>64k tokens allowed), keep being hyper aware to make sure no hallicunated answers(if you need you can use all the self promoting logic organizations etc(like explicitly telling your logic flow to embed in tokens) since hallicunated answers means you don't even give logical answers let alone getting target information, especially considering this is vanilla js, so syntax errors are banned), you can think silently extremely long, triple check everything, as we said it's tricky and this is a test\n\n1)what it does(exclude nothing)\n2)how it does( including things like logic flow, scenario based response, compute requirements, garbage situation, memory usage, self healing, error resistance, statistic based properties, accessing data back etc etc, so exclude nothing)\n3)assume user somehow written wrong c or f values, can data be recovered and how (analyse error rich cases like more than 1 wrong c per f)(it's also possible user forgot to update initialization array after adding new calls)\n\nhere is js code \n\n```JavaScript\nlog(1,7);var debuglogs;function log(f,c){let s;c==void 0&&!isNaN(f)&&([f,c]=(''+f).split('.').map(Number));(!debuglogs||!debuglogs[\"-1\"])&&(s=2,debuglogs=Array(1997).fill().map(()=>[]),debuglogs[\"-1\"]=[2],debuglogs[\"-2\"]=[],debuglogs[0]=[0],debuglogs[999]=[0],[63,5,...Array(98).fill(0),2,2,13,1,1,6,2,1,3,2,1,2,1,1,2,2].forEach((v,i)=>{debuglogs[0][0]+=v,debuglogs[i+1]=[v,...Array(v).fill(0)],debuglogs[1001+i]=[0],debuglogs[999].push(0),s+=v,debuglogs[\"-1\"].push(s)}),debuglogs[998]=[0,1,...Array(debuglogs[0][0]).fill(1)],debuglogs[998][0]=debuglogs[998].filter(x=>!x).length-1,debuglogs[1000]=[0],log(1,1));if(debuglogs[998][1]&&(s=debuglogs[\"-1\"][f-1]+c-1,debuglogs[998][0]==0||debuglogs[998][s])){debuglogs[\"-2\"][s]||(debuglogs[\"-2\"][s]=f+'.'+c,debuglogs[0].push(s),(debuglogs[0].length>debuglogs[0][0]+3||debuglogs[1000].length>1003||debuglogs[1000+f].length>103)&&[[0,'Sequential',debuglogs[0][0]],[1000,'Recent',1000],[1000+f,`Function ${f}`,100]].forEach(([a,t,l])=>debuglogs[a].length>l+3&&alert(`${t} overflow: ${debuglogs[a].length}, limit: ${l+3}`)));debuglogs[f][c]++;debuglogs[999][f]=++debuglogs[1000+f][0];debuglogs[1000][0]=++debuglogs[999][0];debuglogs[1000][(debuglogs[1000][0]%1000)+1]=debuglogs[\"-2\"][s];debuglogs[1000+f][(debuglogs[1000+f][0]%100)+1]=c}};log(1,8);\n```\n" } ] } ] }, { "name": "temp", "type": "folder", "size": "97,124 bytes", "path": "scratchpad/temp", "children": [ { "name": "out", "type": "folder", "size": "97,124 bytes", "path": "scratchpad/temp/out", "children": [ { "name": "html", "type": "folder", "size": "97,124 bytes", "path": "scratchpad/temp/out/html", "children": [ { "name": "0079.html", "type": "file", "size": "54,082 bytes", "path": "scratchpad/temp/out/html/0079.html", "content": "\n\n\n" }, { "name": "communicationhelpertrversion.html", "type": "file", "size": "3,673 bytes", "path": "scratchpad/temp/out/html/communicationhelpertrversion.html", "content": "\n\nİletişim Yardımcısı\n\n" }, { "name": "jitlogtest.html", "type": "file", "size": "33,424 bytes", "path": "scratchpad/temp/out/html/jitlogtest.html", "content": "\n\n\n \n JIT Benchmarker\n \n\n\n

JIT Benchmarker

\n

Select versions to test, or use the execution box to add/modify contestants.

\n
\n
\n Versions to Test\n
\n
\n
\n
\n
\n Dynamic Code Execution (Pure Eval)\n
\n \n \n \n \n
\n
\n
\n
\n
\n Test Scenario\n
\n Predictable Patterns:\n \n \n
\n
\n Randomized Patterns:\n \n \n
\n
\n
\n
\n
\n Reset Strategy\n
\n \n \n
\n
\n
\n
\n
\n \n \n
\n
\n
\n

History

\n
    \n
    \n

    Ready.

    \n\n\n\n\n\n" }, { "name": "primeexp.html", "type": "file", "size": "5,945 bytes", "path": "scratchpad/temp/out/html/primeexp.html", "content": "Prime‑Derivative Match Matrix Explorer\n" } ] } ] } ] }, { "name": "1.cpp", "type": "file", "size": "795 bytes", "path": "scratchpad/1.cpp", "content": "#include\n#include\nunsigned int v[300000];unsigned long long P[300000],B[300000],c=0,q,d,V,C,i,n,k,s,t=0;int main(){double t0=clock(),t1,t2;for(i=3;i<16777216ULL;i+=4){V=i;C=16777216ULL;q=0;d=0;while(q<24){if(V&1){V=V*3+1;C*=3;}else{V>>=1;C>>=1;q++;}if(C<16777216ULL&&(6ULL*C+V<100663296ULL+i)){d=1;break;}}if(!d){v[c]=i;P[c]=C;B[c++]=V;}}t1=clock();for(i=3;i<100663296ULL;i+=4){n=i;while(n>=i){n=n*3+1;n>>=__builtin_ctzll(n);}t++;}for(k=6ULL;k<=596ULL;k++)for(s=0;s10000000000ULL)continue;n=k*P[s]+B[s];n>>=__builtin_ctzll(n);while(n>=i){n=n*3+1;n>>=__builtin_ctzll(n);}t++;}t2=clock();printf(\"Tested %llu numbers\\nInit filter %%: %.3f%%\\nInit time: %.3fs | Total time: %.3fs\\n\",t,100.0-(c/167772.16),(t1-t0)/CLOCKS_PER_SEC,(t2-t0)/CLOCKS_PER_SEC);}\n" }, { "name": "1.js", "type": "file", "size": "4,438 bytes", "path": "scratchpad/1.js", "content": "var configuration={dataset_text:\"emma\\nolivia\\nava\\nisabella\\nsophia\\ncharlotte\\nmia\\namelia\\nharper\\nevelyn\\nabigail\\nemily\\nella\\nelizabeth\\ncamila\\nluna\\nsofia\\navery\\nmila\\naria\",embedding_dim:16,num_heads:4,num_layers:1,block_size:8,max_steps:500,learning_rate:1e-2,temperature:0.5,linearly_separated_step_samples:10,logarithmically_separated_step_samples:10,output_head_of_all_steps:10,output_tail_of_all_steps:5,print_me_after_run_loss_graph_array:[]};((C)=>{let G=C.print_me_after_run_loss_graph_array;G.push([]);let LH=G[G.length-1];let LG=new Set();for(let i=0;iMath.sqrt(-2*Math.log(Math.random()))*Math.cos(2*Math.PI*Math.random()),docs=C.dataset_text.trim().split('\\n').map(x=>x.trim()).filter(x=>x);for(let i=docs.length-1;i>0;i--){let j=Math.floor(Math.random()*(i+1));[docs[i],docs[j]]=[docs[j],docs[i]]}console.log(`num docs: ${docs.length}`);let U=[...new Set(docs.join(''))].sort(),ST=c=>U.indexOf(c),IT=i=>U[i],VS=U.length+1,BOS=VS-1;console.log(`vocab size: ${VS}`);let V=(d,c=[])=>{let n={d,g:0,c,op:()=>{},mt:0,vt:0};n.a=o=>{o=o.d!=null?o:V(o);let r=V(n.d+o.d,[n,o]);r.op=()=>{n.g+=r.g;o.g+=r.g};return r};n.m=o=>{o=o.d!=null?o:V(o);let r=V(n.d*o.d,[n,o]);r.op=()=>{n.g+=o.d*r.g;o.g+=n.d*r.g};return r};n.p=e=>{let r=V(n.d**e,[n]);r.op=()=>{n.g+=e*n.d**(e-1)*r.g};return r};n.l=()=>{let r=V(Math.log(n.d),[n]);r.op=()=>{n.g+=r.g/n.d};return r};n.e=()=>{let r=V(Math.exp(n.d),[n]);r.op=()=>{n.g+=r.d*r.g};return r};n.r=()=>{let r=V(n.d<0?0:n.d,[n]);r.op=()=>{n.g+=(n.d>0)*r.g};return r};return n},P=[],M=(r,c,s=0.02)=>Array(r).fill().map(()=>Array(c).fill().map(()=>(P.push(v=V(s?RN()*s:0))&&v))),DT=(x,w)=>w.map(r=>r.reduce((a,b,i)=>a.a(b.m(x[i])),V(0))),RM=x=>x.map(v=>v.m(x.reduce((a,b)=>a.a(b.p(2)),V(0)).m(x.length**-1).a(1e-5).p(-.5))),SM=x=>{let m=x.reduce((a,b)=>a>b.d?a:b.d,-1e9),e=x.map(v=>v.a(-m).e()),s=e.reduce((a,b)=>a.a(b),V(0));return e.map(v=>v.m(s.p(-1)))},W={te:M(VS,C.embedding_dim),pe:M(C.block_size,C.embedding_dim),lm:M(VS,C.embedding_dim),l:[]};for(let i=0;i{let x=RM(W.te[t].map((z,i)=>z.a(W.pe[p][i])));for(let i=0;ikp.slice(s,s+hd).reduce((a,b,j)=>a.a(b.m(q[s+j])),V(0)).m(hd**-.5));SM(sc).map((w,idx)=>v[i][idx].slice(s,s+hd).map((u,j)=>att[s+j]=(att[s+j]||V(0)).a(u.m(w))))}x=DT(att,L.o).map((u,j)=>u.a(r1[j]));r1=x;x=RM(x);x=DT(DT(x,L.f1).map(u=>u.r().p(2)),L.f2).map((u,j)=>u.a(r1[j]))}return DT(x,W.lm)};for(let s=0;s[]),vs=W.l.map(()=>[]),loss=V(0);for(let i=0;i{let o=[],s=new Set(),f=x=>{if(!s.has(x)){s.add(x);(x.c||[]).map(f);o.push(x)}};f(n);n.g=1;o.reverse().map(x=>x.op())};P.map(p=>p.g=0);BK(avg);let lr=C.learning_rate*0.5*(1+Math.cos(Math.PI*s/C.max_steps));P.map(p=>{p.mt=.9*p.mt+.1*p.g;p.vt=.95*p.vt+.05*p.g**2;let mh=p.mt/(1-.9**(s+1)),vh=p.vt/(1-.95**(s+1));p.d-=lr*mh/(Math.sqrt(vh)+1e-8)});if(LG.has(s))console.log(`step ${String(s+1).padStart(4)} / ${String(C.max_steps).padStart(4)} | loss ${avg.d.toFixed(4)}`)}console.log(\"\\n--- inference ---\");for(let i=0;i<20;i++){let ks=W.l.map(()=>[]),vs=W.l.map(()=>[]),t=BOS,out=\"\";for(let j=0;jx.m(1/C.temperature)),pr=SM(lg),r=Math.random(),acc=0;for(let z=0;z=r){t=z;break}if(t==BOS)break;out+=IT(t)}console.log(`sample ${String(i+1).padStart(2)}: ${out}`)}})(configuration);\n" } ] }, { "name": "shell", "type": "folder", "size": "96,794 bytes", "path": "shell", "children": [ { "name": "old", "type": "folder", "size": "57,220 bytes", "path": "shell/old", "children": [ { "name": "3old1.sh", "type": "file", "size": "869 bytes", "path": "shell/old/3old1.sh", "content": "REDACTED" }, { "name": "3old2.sh", "type": "file", "size": "3,128 bytes", "path": "shell/old/3old2.sh", "content": "REDACTED" }, { "name": "3old3.sh", "type": "file", "size": "5,788 bytes", "path": "shell/old/3old3.sh", "content": "REDACTED" }, { "name": "5old1.sh", "type": "file", "size": "879 bytes", "path": "shell/old/5old1.sh", "content": "REDACTED" }, { "name": "6old1.sh", "type": "file", "size": "16,071 bytes", "path": "shell/old/6old1.sh", "content": "REDACTED" }, { "name": "6old2.sh", "type": "file", "size": "30,485 bytes", "path": "shell/old/6old2.sh", "content": "REDACTED" } ] }, { "name": "1.sh", "type": "file", "size": "7,767 bytes", "path": "shell/1.sh", "content": "{\nSOURCE_DIR_DEFAULT=\"$HOME/storage/downloads/\"; DEST_DIR_DEFAULT=\"/storage/9C33-6BBD/dosya taşıma/1. alan/asıl/Download/\"\n\nlister_formatter='{size=$5; split(\"B K M G T P\",u,\" \"); i=1; while(size>=1000&&i<7){size/=1000;i++} size_str=sprintf(\"%.2f%s\",size,u[i]); filename=\"\"; for(j=9; j<=NF; j++) filename=filename (j>9 ? \" \" : \"\") $j; printf \"%-11s %s\\n\", size_str, filename}'\nget_dir_stats() { local path=\"$1\" folders=0 files=0; if [ -d \"$path\" ]; then folders=$(find \"$path\" -mindepth 1 -maxdepth 1 -type d | wc -l); files=$(find \"$path\" -mindepth 1 -maxdepth 1 -type f | wc -l); fi; echo \"$folders $files\"; }\ninteractive_lister() { local list_path=\"$SOURCE_DIR\" list_head=10 list_tail_start=1; while true; do echo -e \"\\n\\033[1;34m--- Interactive Directory Lister ---\\033[0m\"; echo -e \"Current Settings: Path: \\033[1;37m${list_path}\\033[0m | Slice: \\033[1;37m${list_tail_start}-${list_head}\\033[0m largest files\"; read -p \"$(echo -e \"\\033[1;33mActions: (L)ist, (P)ath, (H)ead, (T)ail Start, (R)eset, (Q)uit > \\033[0m\")\" -r choice; choice=${choice,,}; case \"$choice\" in l) echo -e \"\\nListing files for '${list_path}'...\"; if ! [ -d \"$list_path\" ]; then echo -e \"\\033[1;31mError: Directory not found.\\033[0m\"; continue; fi; local tail_count=$((list_head - list_tail_start + 1)); if [ $tail_count -lt 1 ]; then echo -e \"\\033[1;31mError: Tail Start cannot be greater than Head.\\033[0m\"; continue; fi; ls -lS \"${list_path}\" | grep \"^-\" | head -n \"$list_head\" | tail -n \"$tail_count\" | awk \"$lister_formatter\" || echo \"No files found or error in listing.\";; p) read -p \"Set path: (S)ource, (T)arget, or enter custom path > \" -r path_choice; case \"${path_choice,,}\" in s) list_path=\"$SOURCE_DIR\";; t) list_path=\"$DEST_DIR\";; *) list_path=\"$path_choice\";; esac;; h) read -p \"Show top N largest files (Head) [current: $list_head]: \" list_head;; t) read -p \"Start from file rank (Tail Start) [current: $list_tail_start]: \" list_tail_start;; r) list_path=\"$SOURCE_DIR\"; list_head=10; list_tail_start=1; echo \"Settings reset to defaults.\";; q) return;; *) echo -e \"\\033[1;31mInvalid option.\\033[0m\";; esac; done; }\nlist_category() { local title=$1 color_code=$2; declare -n arr=$3; if [ ${#arr[@]} -gt 0 ]; then echo -e \"\\n\\033[1;${color_code}m--- $title (${#arr[@]}) ---\\033[0m\"; for f in \"${!arr[@]}\"; do if [[ \"$3\" == \"sync_pairs\" ]]; then echo \" '${f}' -> '${arr[$f]}'\"; else echo \" - $f\"; fi; done; fi; }\n\nrun_analysis() {\n local SOURCE_DIR=\"${1%/}/\"; local DEST_DIR=\"${2%/}/\"; echo -e \"\\033[1;34m--- Analyzing Directories ---\\033[0m\"; echo -e \"Source: \\033[1;37m$SOURCE_DIR\\033[0m\"; echo -e \"Dest: \\033[1;37m$DEST_DIR\\033[0m\"; mapfile -t SOURCE_FILES_TO_CHECK < <(find \"$SOURCE_DIR\" -maxdepth 1 -type f -printf \"%f\\n\" 2>/dev/null || true); local TOTAL_TO_CHECK=${#SOURCE_FILES_TO_CHECK[@]}; declare -A sync_pairs not_found_list ambiguous_list; local unpaired_pass1=()\n\n echo -e \"\\n\\033[1;36mPass 1: Checking for exact matches ($TOTAL_TO_CHECK files)...\\033[0m\"; for source_f in \"${SOURCE_FILES_TO_CHECK[@]}\"; do if [ -f \"${DEST_DIR}${source_f}\" ]; then sync_pairs[\"$source_f\"]=\"$source_f\"; echo -e \" - (${#sync_pairs[@]}/$TOTAL_TO_CHECK) \\033[1;32mExact Match:\\033[0m '${source_f}'\"; else unpaired_pass1+=(\"$source_f\"); fi; done; echo \"Found ${#sync_pairs[@]} exact matches. ${#unpaired_pass1[@]} files remain.\"\n\n if [ ${#unpaired_pass1[@]} -gt 0 ]; then read -p \"$(echo -e \"\\033[1;33mAction: Run (F)uzzy search, (S)kip, (I)nject code > \\033[0m\")\" -n 1 -r choice; echo; choice=${choice,,}; case \"$choice\" in f) echo -e \"\\n\\033[1;36mPass 2: Performing fuzzy search...\\033[0m\"; declare -A fuzzy_map; for f in \"${unpaired_pass1[@]}\"; do sanitized_f=$(echo \"$f\" | sed 's/[^a-zA-Z0-9]//g'); fuzzy_map[\"$sanitized_f\"]=\"$f\"; done; while IFS= read -r dest_f; do sanitized_dest=$(echo \"$dest_f\" | sed 's/[^a-zA-Z0-9]//g'); if [[ -n \"${fuzzy_map[$sanitized_dest]}\" ]]; then source_f=\"${fuzzy_map[$sanitized_dest]}\"; source_size=$(stat -c %s \"${SOURCE_DIR}${source_f}\"); dest_size=$(stat -c %s \"${DEST_DIR}${dest_f}\"); if [[ \"$source_size\" == \"$dest_size\" ]]; then sync_pairs[\"$source_f\"]=\"$dest_f\"; echo -e \" - (${#sync_pairs[@]}/$TOTAL_TO_CHECK) \\033[1;32mFuzzy Match:\\033[0m '${source_f}' -> '${dest_f}'\"; else ambiguous_list[\"$source_f\"]=1; fi; fi; done < <(find \"$DEST_DIR\" -maxdepth 1 -type f -printf \"%f\\n\" 2>/dev/null || true);; i) echo -e \"\\033[1;31mWARNING: Entering interactive shell. Type 'exit' to return.\\033[0m\"; bash;; *) echo \"Skipping fuzzy search.\";; esac; fi\n\n for f in \"${SOURCE_FILES_TO_CHECK[@]}\"; do if [[ -z \"${sync_pairs[$f]}\" && -z \"${ambiguous_list[$f]}\" ]]; then not_found_list[\"$f\"]=1; fi; done\n\n echo -e \"\\n\\033[1;35m--- Analysis Complete: Main Menu ---\\033[0m\"; while true; do read -p \"$(echo -e \"\\033[1;33mAction? M:${#sync_pairs[@]} N:${#not_found_list[@]} | (D)irs (L)ist (R)sync (V)erify (S)ubfolders (Q)uit (H)elp > \\033[0m\")\" -r choice; choice=${choice,,}; case \"$choice\" in d) interactive_lister;; l) list_category \"Confirmed Pairs\" \"32\" sync_pairs; list_category \"Not Found\" \"31\" not_found_list; list_category \"Ambiguous\" \"33\" ambiguous_list;; r) if [ ${#not_found_list[@]} -eq 0 ]; then echo \"No 'Not Found' files to rsync.\"; else rsync -av --progress --files-from=<(printf \"%s\\n\" \"${!not_found_list[@]}\") \"$SOURCE_DIR\" \"$DEST_DIR\"; fi;; v) if [ ${#sync_pairs[@]} -eq 0 ]; then echo \"No matched pairs to verify.\"; elif ! command -v sha512sum &> /dev/null; then echo \"Error: 'sha512sum' command not found.\"; else echo -e \"\\033[1;33mVerifying ${#sync_pairs[@]} pairs (SHA512)...\\033[0m\"; for f in \"${!sync_pairs[@]}\"; do echo -n \"Verifying '$f'... \"; H1=$(sha512sum \"${SOURCE_DIR}$f\" | cut -d' ' -f1); H2=$(sha512sum \"${DEST_DIR}${sync_pairs[$f]}\" | cut -d' ' -f1); if [[ \"$H1\" == \"$H2\" ]]; then echo -e \"\\033[1;32mOK\\033[0m\"; else echo -e \"\\033[1;31mFAIL\\033[0m\"; fi; done; echo \"Verification complete.\"; fi;; s) mapfile -t subdirs < <( (find \"$SOURCE_DIR\" -mindepth 1 -maxdepth 1 -type d -printf \"%f\\n\" 2>/dev/null; find \"$DEST_DIR\" -mindepth 1 -maxdepth 1 -type d -printf \"%f\\n\" 2>/dev/null) | sed 's/^[ \\t]*//;s/[ \\t]*$//' | sort -u ); if [ ${#subdirs[@]} -eq 0 ]; then echo \"No subfolders found.\"; else echo -e \"\\nCalculating content stats...\"; local total_folders=${#subdirs[@]}; echo -e \"\\033[1;33mFormat (#/${total_folders}): Status, Contents (F:S-D, f:S-D), Name\\033[0m\"; i=1; for dir in \"${subdirs[@]}\"; do read src_folders src_files < <(get_dir_stats \"${SOURCE_DIR}${dir}\"); read dst_folders dst_files < <(get_dir_stats \"${DEST_DIR}${dir}\"); status=\"\"; if [ -d \"${SOURCE_DIR}${dir}\" ] && [ -d \"${DEST_DIR}${dir}\" ]; then status=\"\\033[1;32m[B]\\033[0m\"; elif [ -d \"${SOURCE_DIR}${dir}\" ]; then status=\"\\033[1;33m[S]\\033[0m\"; else status=\"\\033[1;36m[D]\\033[0m\"; fi; echo -e \"($i/$total_folders) $status F:${src_folders}-${dst_folders} f:${src_files}-${dst_files} $dir\"; ((i++)); done; read -p \"Select #, type name, or (C)ancel: \" sel; if [[ ${sel,,} != \"c\" && -n \"$sel\" ]]; then if [[ $sel =~ ^[0-9]+$ ]] && [ $sel -ge 1 ] && [ $sel -le ${#subdirs[@]} ]; then sel=\"${subdirs[$((sel-1))]}\"; fi; echo -e \"\\n\\033[1;31m--- Descending into: '$sel' ---\\033[0m\\n\"; run_analysis \"${SOURCE_DIR}${sel}\" \"${DEST_DIR}${sel}\"; echo -e \"\\n\\033[1;31m--- Returning to parent menu ---\\033[0m\"; fi; fi;; q) echo \"Exiting current menu.\"; return;; h) echo -e \" (D)irs\\t- Interactive directory lister.\"; echo -e \" (L)ist\\t- Show files by category.\"; echo -e \" (R)sync\\t- Copy 'Not Found' files.\"; echo -e \" (V)erify\\t- SHA512 check matched pairs.\"; echo -e \" (S)ubfolders\\t- Analyze a subfolder.\"; echo -e \" (Q)uit\\t- Exit current menu.\";; *) echo -e \"\\033[1;31mInvalid option. Press (H) for help.\\033[0m\";; esac; done\n}\n\nrun_analysis \"$SOURCE_DIR_DEFAULT\" \"$DEST_DIR_DEFAULT\"\necho -e \"\\n\\033[1;32mScript finished.\\033[0m\"\n\n}\n" }, { "name": "2.sh", "type": "file", "size": "1,461 bytes", "path": "shell/2.sh", "content": "statcomp() { if [[ \"$#\" -lt 2 || \"$#\" -gt 4 ]]; then echo -e \"Usage: statcomp [ip1_or_alias] [ip2_or_alias]\\nDefine aliases with: export myphone=\\\"127.0.0.1:5555\\\"\" >&2; return 1; fi; local C=\"\\033[31m\"; local G=\"\\033[32m\"; local R=\"\\033[0m\"; _sc_get_stat_data() { local file=\"$1\"; local target=\"$2\"; local P_FMT=\"%n\\n%s\\n%b\\n%B\\n%F\\n%d\\n%i\\n%h\\n%a\\n%A\\n%u\\n%U\\n%g\\n%G\\n%C\\n%x\\n%y\\n%z\\n%w\\n\"; if [[ -z \"$target\" ]]; then stat --printf=\"$P_FMT\" \"$file\"; else local ip; if [[ \"$target\" == *.* || \"$target\" == *:* ]]; then ip=\"$target\"; else ip=\"${!target:-$target}\"; fi; adb -s \"$ip\" shell \"stat -c \\$'$P_FMT' '$file'\"; fi; }; paste <(_sc_get_stat_data \"$1\" \"$3\") <(_sc_get_stat_data \"$2\" \"$4\") | awk -F'\\t' -v C=\"$C\" -v G=\"$G\" -v R=\"$R\" '{v1[NR]=$1;v2[NR]=$2}END{split(\"1 6 7\",t);for(k in t)N[t[k]]=1;split(\"9 11 13\",t);for(k in t)P[t[k]]=1;split(\"2 9 16 17 18 19\",t);for(k in t)L[t[k]]=1;for(i=1;i<=NR;i++){if(P[i]){m=v1[i]==v2[i]&&v1[i+1]==v2[i+1];s[i]=\"(\"v1[i]\"/\"v1[i+1]\")\"(m?\"🟢\":\"🔴\")(L[i]?(m?G:C):\"\")\"(\"v2[i]\"/\"v2[i+1]\")\"(L[i]?R:\"\");i++}else{e=N[i]?\"⚪\":v1[i]==v2[i]?\"🟢\":\"🔴\";s[i]=v1[i] e \"\" (L[i]?(e==\"🟢\"?G:C):\"\") v2[i] \"\" (L[i]?R:\"\")}}printf(\" File: %s\\n Size: %s Blocks: %s IO Block: %s %s\\nDevice: %s Inode: %s Links: %s\\nAccess: %s Uid: %s Gid: %s\\nContext: %s\\nAccess: %s\\nModify: %s\\nChange: %s\\n Birth: %s\\n\",s[1],s[2],s[3],s[4],s[5],s[6],s[7],s[8],s[9],s[11],s[13],s[15],s[16],s[17],s[18],s[19])}'; };statcomp;\n" }, { "name": "3.sh", "type": "file", "size": "6,614 bytes", "path": "shell/3.sh", "content": "richcopy() { local S=\"$1\" D=\"$2\" I=\"$3\" P=\"${4:-8022}\" L=\"$5\" PW=\"$6\" SAFE=\"${7:-0}\" CF=\"${8:-0}\" VI=\"${9:-0}\" BUF=\"${10:-5}\" PT=\"${11:-0}\" HD=\"${12}\" TL=\"${13}\"; [ \"$#\" -lt 2 ] && { echo -e \"\\033[1;33mUsage: richcopy [IP] [PORT] [LIST] [PASS] [SAFE] [CREATE] [VERIFY_INT] [BUF_SIZE] [PATCH] [HEAD] [TAIL]\\n\\nNote: For statcomp use https://raw.githubusercontent.com/halitziyakartal/test11/main/shell/2.sh\\033[0m\"; return 1; }; unset T E N ID FC SC SF CHK; local T=0 E=0 N=0 ID=\"N\" FC=0 SF=0 CHK=0; local SC=\"ssh -p $P\" WL=\".work_list\"; local LAST_UI_TIME=0 UI_DELAY=30000 NOW=0; local UI_REF=100; ((VI > 0)) && UI_REF=$VI; trap 'echo -e \"\\n\\033[1;31m❌ Interrupted.\\033[0m\"; rm \"$WL\" \".rc_remote_list.txt\" 2>/dev/null; return 1' INT; echo -e \"\\n\\033[1;34m--- ANALYZING ---\\033[0m\"; [ -n \"$PW\" ] && command -v sshpass >/dev/null && SC=\"sshpass -p '$PW' ssh -p $P\"; if [ -n \"$I\" ]; then ID=$(adb -s $I shell \"[ -d '$S' ] && echo Y\"); else [ -d \"$S\" ] && ID=\"Y\"; fi; apply_filter() { local cmd=\"cat\"; [ -n \"$HD\" ] && cmd=\"$cmd | head -n $HD\"; [ -n \"$TL\" ] && cmd=\"$cmd | tail -n $TL\"; eval \"$cmd\"; }; if [ -n \"$L\" ]; then if [ ! -f \"$L\" ]; then if [ -n \"$I\" ] && [ \"$(adb -s $I shell \"[ -f '$L' ] && echo Y\")\" == \"Y\" ]; then echo \"⬇️ Fetching remote list...\"; adb -s $I pull \"$L\" \".rc_remote_list.txt\" >/dev/null; L=\".rc_remote_list.txt\"; else echo \"❌ List missing: $L\"; return 1; fi; fi; local RAW_T=$(awk 'END {print NR}' \"$L\"); cat \"$L\" | apply_filter > \"$WL\"; sed -i 's/\\r$//' \"$WL\"; echo \"\" >> \"$WL\"; local LIST_T=$(grep -cve '^\\s*$' \"$WL\"); print_bar() { echo -ne \"List: $LIST_T/$RAW_T | Total: \\033[1;36m$SF/$CHK\\033[0m | Exists: \\033[1;33m$E\\033[0m | New: \\033[1;32m$((SF-E))\\033[0m\\r\"; }; while read -r f; do [ -z \"$f\" ] && continue; ((CHK++)); local EXISTS=0; if [[ \"$f\" == *.* ]]; then EXISTS=1; elif [ -n \"$I\" ]; then [ \"$(adb -s $I shell \"[ -d '$S/$f' ] && echo Y\")\" == \"Y\" ] && EXISTS=1; else [ -d \"$S/$f\" ] && EXISTS=1; fi; if ((EXISTS)); then ((SF++)); [ -e \"$D/$f\" ] && ((E++)); fi; NOW=${EPOCHREALTIME/./}; if (( NOW - LAST_UI_TIME >= UI_DELAY )); then print_bar; LAST_UI_TIME=$NOW; fi; done < \"$WL\"; print_bar; echo \"\"; T=$SF; elif [ -n \"$I\" ]; then if [ \"$ID\" == \"Y\" ]; then local RL=$(adb -s $I shell \"cd '$S' && find . -type f\" | apply_filter); T=$(echo \"$RL\" | wc -l); E=$(echo \"$RL\" | while read -r f; do [ -e \"$D/${f#./}\" ] && echo 1; done | wc -l); else T=1; [ -e \"$D/$(basename \"$S\")\" ] && E=1 || E=0; fi; echo -e \"Total: $T | Exists: \\033[1;33m$E\\033[0m | New: \\033[1;32m$((T-E))\\033[0m\"; else if [ \"$ID\" == \"Y\" ]; then T=$(find \"$S\" -type f | apply_filter | wc -l); E=\"?\"; else T=1; [ -e \"$D/$(basename \"$S\")\" ] && E=1 || E=0; fi; echo -e \"Total: $T | Exists: \\033[1;33m$E\\033[0m | New: \\033[1;32m$((T-E))\\033[0m\"; fi; local ASK=1; if [[ \"$SAFE\" == \"y\" || \"$SAFE\" == \"Y\" ]]; then ASK=0; elif [[ \"$SAFE\" =~ ^[0-9]+$ ]] && ((SAFE > 0 && T <= SAFE)); then ASK=0; fi; if ((ASK)); then read -p \"Proceed? [y/N] \" c; [[ \"$c\" != \"y\" ]] && rm \"$WL\" \".rc_remote_list.txt\" 2>/dev/null && return 1; fi; if [ ! -d \"$D\" ]; then if [[ \"$CF\" != \"y\" && \"$CF\" != \"Y\" ]]; then read -p \"Create dest? [y/N] \" c; [[ \"$c\" != \"y\" ]] && rm \"$WL\" \".rc_remote_list.txt\" 2>/dev/null && return 1; fi; mkdir -p \"$D\"; fi; local TIMESTAMP=$(date +\"%Y%m%d%H%M%S\"); local LOG=\"/storage/emulated/0/Download/richcopy_log_${TIMESTAMP}.txt\"; > \"$LOG\"; echo \"Log: $LOG\"; if [[ \"$PT\" != \"y\" && \"$PT\" != \"Y\" && \"$PT\" != \"metadata\" && \"$PT\" != \"patch\" ]]; then echo \"[1/2] Syncing...\"; if [ -z \"$I\" ]; then if [ \"$ID\" == \"Y\" ]; then cp -av \"${S%/}/.\" \"$D/\"; else cp -av \"$S\" \"$D/\"; fi; else local CMD=(\"rsync\" \"-rh\" \"--info=progress2\" \"--no-times\" \"-s\" \"-e\" \"$SC\"); if [ \"$E\" != \"?\" ] && ((E > 0 && ASK == 1)); then read -p \"Overwrite $E existing files? [y/N] \" w; [[ \"$w\" != \"y\" ]] && CMD+=(\"--ignore-existing\"); fi; if [ -n \"$L\" ]; then \"${CMD[@]}\" --files-from=\"$WL\" \"$I:$S/\" \"$D/\"; elif [ \"$ID\" == \"Y\" ]; then \"${CMD[@]}\" \"$I:${S%/}/\" \"$D/\"; else \"${CMD[@]}\" \"$I:$S\" \"$D/\"; fi; fi; else echo \"⚠️ Patch Mode: Skipping Data Sync.\"; fi; echo \"[2/2] Metadata (SSH Conveyor)...\"; local PREV_HEIGHT=0; local COLS=$(tput cols); update_dashboard() { local START_LINE=$(grep -n \"^----------------\" \"$LOG\" | tail -n \"$BUF\" | head -n 1 | cut -d: -f1); [ -z \"$START_LINE\" ] && START_LINE=1; local DASHBOARD=$(tail -n \"+$START_LINE\" \"$LOG\"); local NEW_HEIGHT=0; while IFS= read -r line; do local LEN=${#line}; if ((LEN == 0)); then ((NEW_HEIGHT++)); else ((NEW_HEIGHT += (LEN + COLS - 1) / COLS)); fi; done <<< \"$DASHBOARD\"; ((PREV_HEIGHT > 0)) && echo -ne \"\\033[${PREV_HEIGHT}A\"; echo -ne \"\\033[0J\"; echo \"$DASHBOARD\"; PREV_HEIGHT=$NEW_HEIGHT; }; process() { local L_CNT=0 LAST_P=\"\" LAST_F=\"\" LAST_VERIFIED=0; run_verify() { echo \"---------------------------------------------------\" >> \"$LOG\"; echo \"#$1: $2\" >> \"$LOG\"; if [ -n \"$I\" ]; then if [ \"$ID\" == \"Y\" ] || [ -n \"$L\" ]; then statcomp \"$S/${2#./}\" \"$3\" \"$I\" \"127.0.0.1\" < /dev/null >> \"$LOG\"; else statcomp \"$S\" \"$3\" \"$I\" \"127.0.0.1\" < /dev/null >> \"$LOG\"; fi; else statcomp \"$S/${2#./}\" \"$3\" < /dev/null >> \"$LOG\"; fi; update_dashboard; }; while IFS='|' read -r p a m || [ -n \"$p\" ]; do [ -z \"$p\" ] && continue; m=\"${m%$'\\r'}\"; a=\"${a%$'\\r'}\"; f=\"$D/${p#./}\"; [[ -e \"$f\" && \"$m\" == *\"20\"* ]] && { touch -a -d \"$a\" \"$f\"; touch -m -d \"$m\" \"$f\"; }; ((L_CNT++)); NOW=${EPOCHREALTIME/./}; if (( NOW - LAST_UI_TIME >= UI_DELAY )); then echo -ne \"\\rProcessed: $L_CNT / $T\"; LAST_UI_TIME=$NOW; fi; LAST_P=\"$p\"; LAST_F=\"$f\"; if ((VI > 0 && (L_CNT == 1 || L_CNT % VI == 0))); then run_verify \"$L_CNT\" \"$p\" \"$f\"; LAST_VERIFIED=$L_CNT; fi; done; if ((VI > 0 && L_CNT > 0 && L_CNT != LAST_VERIFIED)); then run_verify \"$L_CNT (LAST)\" \"$LAST_P\" \"$LAST_F\"; fi; echo -ne \"\\rProcessed: $L_CNT / $T\"; echo \"COUNT=$L_CNT\" >> \"$LOG\"; }; if [ -z \"$I\" ]; then if [[ \"$PT\" == \"y\" || \"$PT\" == \"Y\" ]]; then find \"$S\" -printf \"%P|%x|%y\\n\" | apply_filter | process; fi; elif [ -n \"$L\" ]; then local R_TMP=\"/storage/emulated/0/Download/rc_list_${TIMESTAMP}.txt\"; adb -s $I push \"$WL\" \"$R_TMP\" >/dev/null; $SC $I \"cd '$S' && while read f; do find \\\"\\$f\\\" -exec stat -c '%n|%x|%y' {} + || true; done < '$R_TMP'; rm '$R_TMP'\" | process; elif [ \"$ID\" == \"Y\" ]; then local SR=$(dirname \"${S%/}\"); local TG=$(basename \"${S%/}\"); $SC $I \"cd '$SR' && find '$TG' -exec stat -c '%n|%x|%y' {} +\" | apply_filter | process; else $SC $I \"stat -c '%n|%x|%y' '$S'\" | apply_filter | process; fi; local FINAL_COUNT=$(grep \"COUNT=\" \"$LOG\" | tail -n 1 | cut -d= -f2); echo -e \"\\n✅ Done. Processed $FINAL_COUNT items.\"; trap - INT; [ -f \"$WL\" ] && rm \"$WL\" \".rc_remote_list.txt\" 2>/dev/null; };richcopy;\n" }, { "name": "4.sh", "type": "file", "size": "1,392 bytes", "path": "shell/4.sh", "content": "statfilelist() { local D=\"$1\" O=\"$2\" SI=\"$3\" DI=\"$4\" L=\"$5\" PR=\"${6:-adb}\" PO=\"${7:-8022}\" C=\"${8:-0}\"; [[ -z \"$O\" ]] && { echo \"Usage: statfilelist [src] [dst] [list] [proto] [port] [confirm]\" >&2; return 1; }; _x() { if [[ -n \"$1\" ]]; then [[ \"$PR\" == \"ssh\" ]] && ssh -p \"$PO\" \"$1\" \"$2\" || adb -s \"$1\" shell \"$2\"; else eval \"$2\"; fi; }; local E=$(_x \"$DI\" \"[ -e \\\"$O\\\" ] && echo 1\"); if [[ \"$E\" == \"1\" ]]; then [[ \"$((C&2))\" -eq 0 ]] && { read -p \"Overwrite '$O'? [y/N] \" -n 1 -r; echo; [[ ! $REPLY =~ ^[Yy]$ ]] && return 1; }; else [[ \"$((C&1))\" -eq 0 ]] && { read -p \"Create '$O'? [y/N] \" -n 1 -r; echo; [[ ! $REPLY =~ ^[Yy]$ ]] && return 1; }; fi; local CMD=\"export LC_ALL=C.UTF-8; T=\\$(mktemp); trap 'rm -f \\$T' EXIT; $([[ -n \"$L\" ]] && echo \"cat \\\"$L\\\"\" || echo \"find \\\"$D\\\" -maxdepth 1 -type f\") | sort > \\$T; TOT=\\$(wc -l < \\$T); echo \\\"Total: \\$TOT\\\" >&2; echo '{\\\"data\\\":[\\\"'; cat \\$T | tr '\\n' '\\0' | xargs -0 stat | awk -v M=\\$TOT '{gsub(/\\\"/, \\\"\\\\\\\\\\\\\\\"\\\")} /^ *File: /{if(N++>0) printf \\\"\\\\\\\",\\\\n\\\\\\\"\\\"; if(N%1000==0) printf \\\"\\\\r\\\\033[1;33mStat: %d/%d\\\\033[0m\\\", N, M > \\\"/dev/stderr\\\"} {print}'; echo '\\\"], \\\"map\\\":{'; awk '{gsub(/\\\"/, \\\"\\\\\\\\\\\\\\\"\\\", \\$0); printf \\\"\\\\\\\"%s\\\\\\\":%d,\\\\n\\\", \\$0, NR-1}' \\$T | sed '\\$ s/,\\$//'; echo '}}'\"; if [[ \"$SI\" == \"$DI\" ]]; then _x \"$SI\" \"$CMD > \\\"$O\\\"\"; else _x \"$SI\" \"$CMD\" | _x \"$DI\" \"cat > \\\"$O\\\"\"; fi; echo -e \"\\nDone.\"; }\n" }, { "name": "5.sh", "type": "file", "size": "1,854 bytes", "path": "shell/5.sh", "content": "deleter() { local S=\"$1\" D=\"$2\" I=\"$3\" P=\"${4:-8022}\" L=\"$5\" H=\"$6\" TL=\"$7\"; [ -z \"$S\" ] && echo \"Using $(pwd)\" || { echo \"CD $S\"; cd \"$S\" || return 1; }; { touch t1 t2 t3 t4; echo -e \"\\033[1;34m--- STEP 1: Slicing List ---\\033[0m\"; head -n \"$H\" \"$L\" | tail -n \"$TL\" | tr -d '\\r' > t1; local TOT=$(awk 'END {print NR}' t1); local CNT=0 FOUND=0 LAST=0 NOW=0; echo -e \"\\033[1;34m--- STEP 2: Local Scan ---\\033[0m\"; printf \"\\r> Scanned: 0 / %d | Found: 0\" \"$TOT\" >&2; while read -r f || [ -n \"$f\" ]; do [ -z \"$f\" ] && continue; if [ -f \"$f\" ]; then echo \"$f\"; ((FOUND++)); fi; ((CNT++)); NOW=${EPOCHREALTIME/./}; if ((CNT==1 || NOW-LAST>=30000)); then printf \"\\r> Scanned: %d / %d | Found: \\033[1;33m%d\\033[0m\" \"$CNT\" \"$TOT\" \"$FOUND\" >&2; LAST=$NOW; fi; done < t1 > t2; printf \"\\r> Local Scan: \\033[1;32mDONE\\033[0m. Found \\033[1;32m%d\\033[0m existing files. \\033[K\\n\" \"$FOUND\" >&2; echo -e \"\\033[1;34m--- STEP 3: Remote Scan ---\\033[0m\"; CNT=0; LAST=0; printf \"\\r> Found Remote: 0\" >&2; cat t1 | ssh -p \"$P\" \"$I\" \"cd \\\"$D\\\" && while read -r r || [ -n \\\"\\$r\\\" ]; do [ -z \\\"\\$r\\\" ] && continue; [ -f \\\"\\$r\\\" ] && echo \\\"\\$r\\\"; done\" | while read -r f || [ -n \"$f\" ]; do [ -z \"$f\" ] && continue; echo \"$f\"; ((CNT++)); NOW=${EPOCHREALTIME/./}; if ((CNT==1 || NOW-LAST>=30000)); then printf \"\\r> Found Remote: \\033[1;32m%d\\033[0m\" \"$CNT\" >&2; LAST=$NOW; fi; done > t3; printf \"\\r> Remote Scan: \\033[1;32mDONE\\033[0m. Found \\033[1;32m%d\\033[0m matches. \\033[K\\n\" \"$CNT\" >&2; grep -Fxf t3 t2 > t4 || true; echo -e \"\\n\\033[1;36m=== SUMMARY ===\\033[0m\\nRequested : $TOT\\nFound Here : $(wc -l < t2)\\nFound Remote: $(wc -l < t3)\\n\\033[1;31mTO DELETE : $(wc -l < t4)\\033[0m (Overlap)\\n----------------\\nFirst: $(head -n 1 t4)\\nLast : $(tail -n 1 t4)\\n\"; read -p \"Delete 'TO DELETE' files? (y/n) \" c; [ \"$c\" = \"y\" ] && xargs -d '\\n' -a t4 rm -v; rm t1 t2 t3 t4; }; }\n" }, { "name": "6.sh", "type": "file", "size": "20,486 bytes", "path": "shell/6.sh", "content": "askai(){ type ai>/dev/null 2>&1||ai(){ askai \"$@\";};local CONF_DIR=\"${XDG_CONFIG_HOME:-$HOME/.config}/askai\" PATH_FILE=\"$CONF_DIR/.path\" D=\"${TMPDIR:-/tmp}/g_ai\" API=\"https://generativelanguage.googleapis.com/v1beta/models\";if [ -s \"$PATH_FILE\" ];then read -r SAVED_D<\"$PATH_FILE\";if [ -d \"$SAVED_D\" ]&&[ -w \"$SAVED_D\" ];then D=\"$SAVED_D\";else echo -e \"\\033[1;33mWarning: Persistent path '$SAVED_D' inaccessible. Falling back to $D\\033[0m\">&2;fi;fi;mkdir -p \"$D\" 2>/dev/null&&chmod 700 \"$D\"||{ echo \"Error: Failed to secure $D\">&2;return 1;};if ! command -v jq>/dev/null||! command -v curl>/dev/null;then echo -e \"\\033[1;31mError: Missing curl/jq.\\033[0m\">&2;return 1;fi;local PF RO ERR_TMP=\"\" pipe_smart=0 MACRO_Q=();PF=$(mktemp \"$D/payload.XXXXXX.json\")||return 1;RO=$(mktemp \"$D/r.XXXXXX.tmp\")||{ rm -f \"$PF\";return 1;};trap 'rm -f \"$PF\" \"$RO\" \"${ERR_TMP}\" \"$D\"/slice.*.json 2>/dev/null;unset -f require_interaction handle_err _rollback_last _display_history _chat_submenu _ui_read _jq_in _slice_hist 2>/dev/null' RETURN;_jq_in(){ local f=\"${!#}\" t;t=$(mktemp \"$D/j.XXXXXX\")||return 1;if jq \"${@:1:$#-1}\" \"$f\">\"$t\" 2>/dev/null&&[ -s \"$t\" ];then cat \"$t\">\"$f\"&&rm -f \"$t\";else rm -f \"$t\";return 1;fi;};_slice_hist(){ local s=\"$1\" d=\"$2\" hn=\"$3\" tn=\"$4\" ht=\"$5\" tt=\"$6\";if [ -n \"$hn\" ];then jq \".[0:${hn}]\" \"$s\">\"$d\";elif [ -n \"$tn\" ];then [ \"$tn\" -eq 0 ]&&echo \"[]\">\"$d\"||jq \".[-${tn}:]\" \"$s\">\"$d\";elif [ -n \"$ht\" ];then jq --argjson ht \"$ht\" '. as $m|reduce range(length) as $i({r:[],t:0};($m[$i].parts[0].text//\"\"|length/4|floor)as $k|if .t+$k<=$ht then{r:(.r+[$m[$i]]),t:(.t+$k)}else . end)|.r' \"$s\">\"$d\";elif [ -n \"$tt\" ];then jq --argjson tt \"$tt\" '. as $m|(length-1)as $l|reduce range($l;-1;-1)as $i({r:[],t:0};($m[$i].parts[0].text//\"\"|length/4|floor)as $k|if .t+$k<=$tt then{r:[$m[$i]]+.r,t:(.t+$k)}else . end)|.r' \"$s\">\"$d\";else cp \"$s\" \"$d\";fi;};_ui_read(){ local is_sec=0;[ \"$1\" == \"-s\" ]&&{ is_sec=1;shift;};local __v=$1 __p=$2 val=\"\";if [ ${#MACRO_Q[@]} -gt 0 ];then val=\"${MACRO_Q[0]}\";MACRO_Q=(\"${MACRO_Q[@]:1}\");[ \"$is_sec\" -eq 1 ]&&echo -e \"${__p}\\033[1;35m***\\033[0m (macro)\">&2||echo -e \"${__p}\\033[1;35m${val}\\033[0m (macro)\">&2;printf -v \"$__v\" \"%s\" \"$val\";return 0;fi;if [ \"$pipe_smart\" -eq 1 ]&&[ ! -t 0 ];then if [ \"$__p\" == \"Prompt > \" ];then val=$(cat);if [ -n \"$val\" ];then echo -e \"${__p}\\033[1;35m(stream read)\\033[0m\">&2;printf -v \"$__v\" \"%s\" \"$val\";return 0;fi;elif IFS= read -r val;then [ \"$is_sec\" -eq 1 ]&&echo -e \"${__p}\\033[1;35m***\\033[0m (smart)\">&2||echo -e \"${__p}\\033[1;35m${val}\\033[0m (smart)\">&2;printf -v \"$__v\" \"%s\" \"$val\";return 0;fi;fi;if [ \"$interactive\" -eq 0 ];then echo -e \"\\n\\033[1;31mError: Interaction required.\\033[0m\">&2;return 1;fi;if [ \"$is_sec\" -eq 1 ];then read -s -p \"$__p\" val&2;else read -p \"$__p\" val&2;return 1;};};_rollback_last(){ _jq_in 'del(.[-1])' \"$HF\";};handle_err(){ echo -e \"\\n\\033[1;31m❌ API Error:\\033[0m $1\">&2;_rollback_last;if [[ \"$1\" == *\"limit: 0\"* ]];then echo -e \"\\033[1;33m⚠️ Quota Exceeded (Limit 0). Blocking '$M'.\\033[0m\">&2;local kh;kh=$(cksum<<<\"$k\"|cut -d' ' -f1);_jq_in --arg k \"$kh\" --arg m \"$M\" '.[$k]+=[$m]|.[$k]|=unique' \"$BF\";fi;};_display_history(){ local hf=\"$1\" skip_tok=\"${2:-0}\" label=\"$3\" hn=\"$4\" tn=\"$5\" ht=\"$6\" tt=\"$7\";if [ ! -s \"$hf\" ]||[ \"$(cat \"$hf\")\" = \"[]\" ];then echo -e \"\\033[1;33m(No history for '$label')\\033[0m\";return 0;fi;local sliced_tmp;sliced_tmp=$(mktemp \"$D/slice.XXXXXX.json\")||return 1;_slice_hist \"$hf\" \"$sliced_tmp\" \"$hn\" \"$tn\" \"$ht\" \"$tt\";local total shown shown_tok;total=$(jq 'length' \"$hf\");shown=$(jq 'length' \"$sliced_tmp\");shown_tok=$(jq '[.[].parts[0].text//\"\"|length]|add//0|./4|floor' \"$sliced_tmp\");echo -e \"\\033[1;34m─── History: $label ───\\033[0m\";if [ -n \"$hn\" ]||[ -n \"$tn\" ]||[ -n \"$ht\" ]||[ -n \"$tt\" ];then echo -e \"\\033[2mShowing $shown of $total messages | ~$shown_tok tokens\\033[0m\";else echo -e \"\\033[2m$shown messages | ~$shown_tok tokens\\033[0m\";fi;echo \"\";jq -r --argjson skip \"$skip_tok\" 'to_entries[]|(.key+1)as $i|.value.role as $r|(.value.parts[0].text//\"\")as $t|($t|length/4|floor)as $k|(if $r==\"user\" then \"\\u001b[1;36m[\\($i)] You (~\\($k) tokens):\\u001b[0m\" else \"\\u001b[1;32m[\\($i)] Gemini (~\\($k) tokens):\\u001b[0m\" end),(if $skip>0 and $k>$skip then \" \\u001b[2m[skipped — ~\\($k) toks (exceeds skip limit)]\\u001b[0m\" else $t end), \"\"' \"$sliced_tmp\" 2>/dev/null;};_chat_submenu(){ local tgt=\"$1\" c_head=\"\" c_tail=\"\" c_ht=\"\" c_tt=\"\" c_skip=\"0\" b_id=\"\" c_sel=\"\" src_hf=\"$D/h_${tgt}.json\";[ -f \"$src_hf\" ]&&jq -e . \"$src_hf\">/dev/null 2>&1||echo \"[]\">\"$src_hf\";while true;do echo -e \"\\n\\033[1;34m--- Chat Menu: $tgt ---\\033[0m\\n1) \\033[1;32mSet as Default & Use\\033[0m\\n2) \\033[1;36mUse Temporarily\\033[0m\\n3) \\033[1;33mView History\\033[0m (Applies filters)\\n4) \\033[1;35mBranch to New ID\\033[0m (Forks using filters)\\n----------------------------------------\\n5) Msg Head limit: \\033[1;36m[${c_head:-(All)}]\\033[0m\\n6) Msg Tail limit: \\033[1;36m[${c_tail:-(All)}]\\033[0m\\n7) Token Head limit: \\033[1;36m[${c_ht:-(All)}]\\033[0m\\n8) Token Tail limit: \\033[1;36m[${c_tt:-(All)}]\\033[0m\\n9) Skip-Over (Tokens): \\033[1;36m[${c_skip:-0}]\\033[0m\\nr) Return to Chat List\\n0) Abort\">&2;_ui_read c_sel \"Select > \"||return 0;case \"$c_sel\" in 1) return 10;;2) return 11;;3) _display_history \"$src_hf\" \"$c_skip\" \"$tgt\" \"$c_head\" \"$c_tail\" \"$c_ht\" \"$c_tt\";_ui_read c_sel \"(Press Enter to continue...) \";;4) _ui_read b_id \"Enter New Branch ID: \";b_id=$(sed 's/[^a-zA-Z0-9_-]//g'<<<\"$b_id\");[ -z \"$b_id\" ]&&continue;local dst_hf=\"$D/h_${b_id}.json\" dst_sf=\"$D/s_${b_id}.txt\" src_sf=\"$D/s_${tgt}.txt\";if [ -f \"$dst_hf\" ];then _ui_read c_sel \"'$b_id' exists. Overwrite? (y/N): \";[[ ! \"$c_sel\" =~ ^[Yy] ]]&&continue;fi;local sl_tmp;sl_tmp=$(mktemp \"$D/slice.XXXXXX.json\");_slice_hist \"$src_hf\" \"$sl_tmp\" \"$c_head\" \"$c_tail\" \"$c_ht\" \"$c_tt\";mv \"$sl_tmp\" \"$dst_hf\";[ -s \"$src_sf\" ]&&cp \"$src_sf\" \"$dst_sf\"||touch \"$dst_sf\";echo -e \"\\033[1;32m✅ Branched to '$b_id'.\\033[0m\">&2;chat_id=\"$b_id\";return 2;;5) _ui_read c_head \"Msg Head (empty for All): \";c_head=$(grep -o '^[0-9]*'<<<\"$c_head\");c_tail=\"\";c_ht=\"\";c_tt=\"\";;6) _ui_read c_tail \"Msg Tail (empty for All): \";c_tail=$(grep -o '^[0-9]*'<<<\"$c_tail\");c_head=\"\";c_ht=\"\";c_tt=\"\";;7) _ui_read c_ht \"Token Head (empty for All): \";c_ht=$(grep -o '^[0-9]*'<<<\"$c_ht\");c_head=\"\";c_tail=\"\";c_tt=\"\";;8) _ui_read c_tt \"Token Tail (empty for All): \";c_tt=$(grep -o '^[0-9]*'<<<\"$c_tt\");c_head=\"\";c_tail=\"\";c_ht=\"\";;9) _ui_read c_skip \"Skip-Over Tokens (0 disable): \";c_skip=$(grep -o '^[0-9]*'<<<\"$c_skip\");;[rR]) return 12;;0) return 0;;esac;done;};local BF=\"$D/b.json\" STF=\"$D/stream.txt\" IDF=\"$D/id.txt\" KF=\"$D/k.txt\" AF=\"$D/a.txt\" MF=\"$D/m.txt\" CF=\"$D/c.txt\" WF=\"$D/w.json\";for f in \"$BF\" \"$WF\";do jq -e . \"$f\">/dev/null 2>&1||echo \"{}\">\"$f\";done;[[ \"$(<\"$STF\" 2>/dev/null)\" != \"1\" ]]&&echo \"0\">\"$STF\";[[ \"$(<\"$CF\" 2>/dev/null)\" != \"1\" ]]&&echo \"0\">\"$CF\";local chat_id=\"default\";local ci=$(sed 's/[^a-zA-Z0-9_-]//g' \"$IDF\" 2>/dev/null);[ -n \"$ci\" ]&&chat_id=\"$ci\"||echo \"default\">\"$IDF\";local M=\"gemini-flash-latest\";[ -s \"$MF\" ]&&M=$(<\"$MF\");local k=\"\" n=0 s=\"\" mo=0 show_all=0 stream_mode=0 select_id=0 del_mode=0 manage_key=0 arg_id=\"\" arg_m=\"\" arg_k=\"\" p=\"\";local interactive=1 has_pipe=0 continuous=0;[ ! -t 0 ]&&has_pipe=1;[ \"$(<\"$CF\" 2>/dev/null)\" == \"1\" ]&&continuous=1;if [ \"$#\" -eq 0 ]&&[ \"$has_pipe\" -eq 0 ];then echo -e \"\\033[1;33mUsage:\\033[0m askai [flags] \\n\\033[1;36mContext:\\033[0m ID=[$chat_id] | Model=[$M]\\n\\033[1;37mSession Flags:\\033[0m\\n -id [chat_name] Select Chat, View History, or Branch (Menu)\\n -n New Chat Menu (Clear/Rename)\\n -d Delete History Menu\\n\\033[1;37mConfig Flags:\\033[0m\\n -m [model_name] Select/Edit Model (empty to list)\\n -k [api_key] Set/Manage API Key\\n -s [sys_prompt] Set System Prompt\\n -c, --continuous Toggle/Set Continuous Mode\\n --stream Toggle/Set Streaming Output\\n --persistence Move/Edit Data Storage Location\\n\\033[1;37mAdvanced / Automation:\\033[0m\\n -pipe smart Automate UI via stdin lines, remainder is prompt\\n -pipe hard Disable interactive menus entirely\\n -id|name|val|val Pipe automation via arguments (Macro string)\">&2;return 0;fi;while [[ \"$#\" -gt 0 ]];do case \"$1\" in -h|--help) askai;return 0;;--persistence) require_interaction \"Persistence Menu\"||return 1;echo -e \"\\033[1;34m--- Persistence ---\\033[0m\\nCurrent: $D\">&2;local np;_ui_read np \"New path: \"||return 1;[ -z \"$np\" ]&&return 0;np=\"${np/#\\~/$HOME}\";if mkdir -p \"$CONF_DIR\" \"$np\" 2>/dev/null&&chmod 700 \"$np\" 2>/dev/null&&touch \"$np/.test\" 2>/dev/null;then rm -f \"$np/.test\";cp -r \"$D/\"* \"$np/\" 2>/dev/null;echo \"$np\">\"$PATH_FILE\";D=\"$np\";echo -e \"\\033[1;32mMigrated to $D\\033[0m\">&2;return 0;else echo -e \"\\033[1;31mError: Path '$np' invalid/unwritable.\\033[0m\">&2;return 1;fi;;-c|--continuous) if [[ \"$#\" -eq 1 && -z \"$p\" ]];then require_interaction \"Continuous Menu\"||return 1;echo -e \"\\033[1;34m--- Continuous Settings ---\\033[0m\\n1) Enable Always\\n2) Enable Once\\n3) Disable Default\\n0) Abort\">&2;local c;_ui_read c \"Choice > \"||return 1;case \"$c\" in 1) echo \"1\">\"$CF\";return 0;;2) continuous=1;return 0;;3) echo \"0\">\"$CF\";return 0;;*) return 0;;esac;else continuous=1;shift;fi;;--stream) if [[ \"$#\" -eq 1 && -z \"$p\" ]];then require_interaction \"Stream Menu\"||return 1;echo -e \"\\033[1;34m--- Stream Settings ---\\033[0m\\n1) Enable Always\\n2) Enable Once\\n0) Abort\">&2;local c;_ui_read c \"Choice > \"||return 1;[ \"$c\" = \"1\" ]&&echo \"1\">\"$STF\";[ \"$c\" = \"2\" ]&&stream_mode=1;return 0;else stream_mode=1;shift;fi;;-pipe) if [[ \"$2\" == \"hard\" ]];then interactive=0;shift 2;elif [[ \"$2\" == \"smart\" ]];then pipe_smart=1;interactive=1;shift 2;else return 1;fi;;-k|--key) if [[ -n \"$2\" && ! \"$2\" =~ ^- ]];then arg_k=\"$2\";shift 2;else manage_key=1;shift;fi;;-id|--id) if [[ -n \"$2\" && ! \"$2\" =~ ^- ]];then if [[ \"$2\" == *\"|\"* ]];then arg_id=\"${2%%|*}\";set -f;IFS='|' read -r -a new_macros<<<\"${2#*|}\";set +f;MACRO_Q+=(\"${new_macros[@]}\");else arg_id=\"$2\";fi;shift 2;else select_id=1;shift;fi;;-id\\|*|--id\\|*) local fv=\"${1#*|}\";arg_id=\"${fv%%|*}\";local m=\"${fv#*|}\";if [[ \"$m\" != \"$fv\" ]];then set -f;IFS='|' read -r -a new_macros<<<\"$m\";set +f;MACRO_Q+=(\"${new_macros[@]}\");fi;shift;;-n) n=1;shift;;-d|--delete) del_mode=1;shift;;-s) if [[ -n \"$2\" ]];then s=\"$2\";shift 2;else return 1;fi;;-m) mo=1;if [[ -n \"$2\" && ! \"$2\" =~ ^- ]];then arg_m=\"$2\";shift 2;else shift;fi;;-1) show_all=1;shift;;*) p=\"${p:+$p }$1\";shift;;esac;done;local is_ui=0;[[ -z \"$p\" && \"$has_pipe\" -eq 0 && \"$pipe_smart\" -eq 0 && ${#MACRO_Q[@]} -eq 0 ]]&&is_ui=1;if [ \"$manage_key\" -eq 1 ];then require_interaction \"Key Manager\"||return 1;local ck=\"None\" ca=\"Off\";[ -s \"$KF\" ]&&ck=\"...$(<\"$KF\"|tail -c 5)\";[ \"$(<\"$AF\" 2>/dev/null)\" = \"1\" ]&&ca=\"On\";echo -e \"\\033[1;33m--- Key Manager ---\\033[0m\\nStored: $ck | Auto-Use: $ca\\n1) Update Key\\n2) Toggle Auto\\n3) Temp Key\\n0) Abort\">&2;local kc;_ui_read kc \"Choice > \"||return 1;case \"$kc\" in 1) local nk;_ui_read -s nk \"New Key: \"||return 1;echo \"$nk\">\"$KF\";echo 1>\"$AF\";return 0;;2) [ \"$ca\" == \"On\" ]&&echo 0>\"$AF\"||echo 1>\"$AF\";return 0;;3) _ui_read -s arg_k \"Temp Key: \"||return 1;;*) return 0;;esac;fi;if [ \"$del_mode\" -eq 1 ];then require_interaction \"Delete Menu\"||return 1;echo -e \"\\033[1;31m--- Delete Menu ---\\033[0m\\n1) Delete ALL histories\\n2) Delete Current [$chat_id]\\n0) Abort\">&2;local dc;_ui_read dc \"Choice > \"||return 1;case \"$dc\" in 1) rm -f \"$D\"/h_*.json;return 0;;2) rm -f \"$D/h_${chat_id}.json\";chat_id=\"default\";echo \"default\">\"$IDF\";return 0;;*) return 0;;esac;fi;if [ \"$n\" -eq 1 ];then require_interaction \"New Chat Menu\"||return 1;while true;do local nid;_ui_read nid \"New Session ID: \"||return 1;local clean_nid=$(sed 's/[^a-zA-Z0-9_-]//g'<<<\"$nid\");[ -z \"$clean_nid\" ]&&continue;if [ -f \"$D/h_${clean_nid}.json\" ];then echo -e \"\\033[1;33mExists.\\033[0m\\n1) Overwrite\\n2) New Name\\n3) Rename Old\\n0) Abort\">&2;local nc;_ui_read nc \"> \"||return 1;case \"$nc\" in 1) chat_id=\"$clean_nid\";rm -f \"$D/h_${chat_id}.json\";echo \"$chat_id\">\"$IDF\";break;;2) continue;;3) local old;_ui_read old \"Rename to: \"||return 1;mv \"$D/h_${clean_nid}.json\" \"$D/h_${old}.json\";chat_id=\"$clean_nid\";echo \"$chat_id\">\"$IDF\";break;;*) return 0;;esac;else chat_id=\"$clean_nid\";echo \"$chat_id\">\"$IDF\";break;fi;done;fi;if [ -n \"$arg_id\" ];then if [ \"$is_ui\" -eq 1 ];then require_interaction \"ID Selector\"||return 1;_chat_submenu \"$arg_id\";local ret=$?;[ $ret -eq 10 ]&&{ chat_id=\"$arg_id\";echo \"$chat_id\">\"$IDF\";};[ $ret -eq 11 ]&&chat_id=\"$arg_id\";[ $ret -eq 12 ]||[ $ret -eq 0 ]&&return 0;else chat_id=\"$arg_id\";fi;fi;if [ \"$select_id\" -eq 1 ];then require_interaction \"Session List\"||return 1;while true;do echo -e \"\\n\\033[1;34m--- Select Chat ---\\033[0m\">&2;local fl=();while IFS= read -r f;do [ -n \"$f\" ]&&fl+=(\"$f\");done< <(ls -t \"$D\"/h_*.json 2>/dev/null);local ids=() i=1;if [ ${#fl[@]} -gt 0 ];then for f in \"${fl[@]}\";do local nm=$(basename \"$f\"|sed 's/^h_//;s/\\.json$//');ids+=(\"$nm\");read -r msg_n chars_n<<<\"$(jq -r '(length|tostring)+\" \"+([.[].parts[].text//\"\"]|join(\"\")|length|tostring)' \"$f\" 2>/dev/null||echo \"0 0\")\";local cur=\"\";[ \"$nm\" == \"$chat_id\" ]&&cur=\"\\033[1;32m(*)\\033[0m\";printf \"%2d) \\033[1;36m%-18s\\033[0m %8d chars | %6d toks | %4d msgs %b\\n\" \"$i\" \"$nm\" \"$chars_n\" \"$((chars_n/4))\" \"$msg_n\" \"$cur\">&2;((i++));done;else echo \"(No histories)\">&2;fi;echo -e \"\\n+) Create New\\n0) Abort\">&2;local sel;_ui_read sel \"Select > \"||return 1;if [[ \"$sel\" == \"0\" ]];then return 0;elif [[ \"$sel\" == \"+\" ]];then local nn;_ui_read nn \"New Name: \"||return 1;local pk=$(sed 's/[^a-zA-Z0-9_-]//g'<<<\"$nn\");[ -z \"$pk\" ]&&continue;_chat_submenu \"$pk\";local ret=$?;[ $ret -eq 10 ]&&{ chat_id=\"$pk\";echo \"$chat_id\">\"$IDF\";break;};[ $ret -eq 11 ]&&{ chat_id=\"$pk\";break;};[ $ret -eq 2 ]&&break;[ $ret -eq 0 ]&&return 0;elif [[ \"$sel\" =~ ^[0-9]+$ && \"$sel\" -le \"${#ids[@]}\" && \"$sel\" -gt 0 ]];then local pk=\"${ids[$((sel-1))]}\";_chat_submenu \"$pk\";local ret=$?;[ $ret -eq 10 ]&&{ chat_id=\"$pk\";echo \"$chat_id\">\"$IDF\";break;};[ $ret -eq 11 ]&&{ chat_id=\"$pk\";break;};[ $ret -eq 2 ]&&break;[ $ret -eq 0 ]&&return 0;fi;done;fi;local HF=\"$D/h_${chat_id}.json\" SF=\"$D/s_${chat_id}.txt\";jq -e . \"$HF\">/dev/null 2>&1||echo \"[]\">\"$HF\";[ ! -f \"$SF\" ]&&touch \"$SF\";if [ -n \"$arg_k\" ];then if [ \"$is_ui\" -eq 1 ];then require_interaction \"Key Confirm\"||return 1;echo -e \"\\033[1;33mAPI Key: '...${arg_k: -4}'\\033[0m\\n1) Save Default\\n2) Use Temp\\n0) Abort\">&2;local c;_ui_read c \"Choice > \"||return 1;case \"$c\" in 1) k=\"$arg_k\";echo \"$k\">\"$KF\";echo 1>\"$AF\";;2) k=\"$arg_k\";;*) return 0;;esac;else k=\"$arg_k\";fi;fi;if [ -z \"$k\" ];then if [ -s \"$KF\" ]&&[ \"$(<\"$AF\" 2>/dev/null)\" = \"1\" ];then k=$(<\"$KF\");else require_interaction \"Setup\"||return 1;_ui_read -s k \"Enter API Key: \"||return 1;echo \"$k\">\"$KF\";echo 1>\"$AF\";fi;fi;if [ -n \"$arg_m\" ];then if [ \"$is_ui\" -eq 1 ];then require_interaction \"Model Confirm\"||return 1;echo -e \"\\033[1;33mModel: '$arg_m'\\033[0m\\n1) Save Default\\n2) Use Temp\\n0) Abort\">&2;local c;_ui_read c \"Choice > \"||return 1;case \"$c\" in 1) M=\"$arg_m\";echo \"$M\">\"$MF\";;2) M=\"$arg_m\";;*) return 0;;esac;else M=\"$arg_m\";fi;fi;if [ \"$mo\" -eq 1 ]&&[ -z \"$arg_m\" ];then require_interaction \"Model List\"||return 1;while true;do echo -e \"\\033[1;30mFetching models...\\033[0m\">&2;local jm;jm=$(curl -s --connect-timeout 10 -H \"x-goog-api-key: $k\" \"$API\");local err=$(jq -r '.error.message//empty' 2>/dev/null<<<\"$jm\");if [ -n \"$err\" ];then echo -e \"\\033[1;31m❌ API Error:\\033[0m $err\">&2;return 1;fi;local bk=$(<\"$BF\") kh=$(cksum<<<\"$k\"|cut -d' ' -f1) jq_s='.models[]|select(.supportedGenerationMethods[]?|contains(\"generateContent\"))|.name|=sub(\"^models/\";\"\")|.score=0|if(.name|contains(\"-pro\"))then .score+=1000 elif(.name|contains(\"-flash\"))then .score+=800 elif(.name|contains(\"-lite\"))then .score+=600 else . end|if(.name|test(\"gemini-[0-9]\\\\.[0-9]\"))then .score+=((.name|capture(\"gemini-(?[0-9]\\\\.[0-9])\").v|tonumber)*100)else . end|if(.name|contains(\"-latest\"))then .score+=10000 elif(.name|contains(\"-exp\"))then .score+=-5000 elif(.name|contains(\"-preview\"))then if(.name|test(\"preview-[0-9]{2}-[0-9]{2}\"))then .score+=-2000 else .score+=500 end else .score+=5000 end|if(.name|test(\"gemma|learnlm|tts|image|robotics|computer-use|thinking\"))then .score+=-50000 else . end|select($a==\"1\" or(.name as $n|($bk[$h]//[])|index($n)|not))|{name:.name,score:.score}';local ml=();mapfile -t ml< <(jq -r --arg h \"$kh\" --arg a \"$show_all\" --argjson bk \"$bk\" \"$jq_s\"<<<\"$jm\"|jq -rs 'sort_by(.score)|reverse|.[].name');local hc=$(jq -r --arg h \"$kh\" '.[$h]//[]|length'<<<\"$bk\") i=1;for md in \"${ml[@]}\";do if [ \"$md\" == \"$M\" ];then echo -e \"$i) \\033[1;32m$md (*)\\033[0m\">&2;else echo \"$i) $md\">&2;fi;((i++));done;echo \"-1) list 0 limit models too ($hc hidden)\">&2;local sl;_ui_read sl \"Select (1-${#ml[@]}, 0 Abort): \"||return 1;if [ \"$sl\" == \"-1\" ];then show_all=$((1-show_all));continue;fi;[ \"$sl\" = \"0\" ]&&return 0;if [[ \"$sl\" =~ ^[0-9]+$ && \"$sl\" -le \"${#ml[@]}\" ]];then local sel=\"${ml[$((sl-1))]}\";echo -e \"\\033[1;33m$sel\\033[0m\\n1) Save Default\\n2) Use Temp\\n3) Cancel\\n4) Retry\">&2;local sc;_ui_read sc \"Choice > \"||return 1;case \"$sc\" in 1) M=\"$sel\";echo \"$M\">\"$MF\";break;;2) M=\"$sel\";break;;3) break;;4) continue;;*) return 0;;esac;fi;done;fi;[ -n \"$s\" ]&&echo \"$s\">\"$SF\";if [ \"$has_pipe\" -eq 1 ]&&[ -z \"$p\" ]&&[ \"$pipe_smart\" -eq 0 ]&&[ ${#MACRO_Q[@]} -eq 0 ];then p=$( \"||return 1;elif [ \"$continuous\" -eq 1 ]||{ [ \"$pipe_smart\" -eq 1 ]&&[ ! -t 0 ];};then echo -e \"\\033[1;34m> (Stream read... Ctrl+D to send)\\033[0m\">&2;p=$(cat);else _ui_read p \"Prompt > \"||return 1;fi;[ -z \"$p\" ]&&return 0;fi;read -r msg_n h_chars<<<\"$(jq -r '(length+1|tostring)+\" \"+([.[].parts[].text//\"\"]|join(\"\")|length|tostring)' \"$HF\" 2>/dev/null||echo \"1 0\")\";if [ \"$(((${h_chars:-0}+${#p})/4))\" -gt 200000 ]&&! jq -e --arg id \"$chat_id\" '.[$id]' \"$WF\">/dev/null 2>&1;then echo -e \"\\033[1;33m⚠️ Session '$chat_id' exceeded ~200k tokens. Trim via askai -id > Branch\\033[0m\">&2;_jq_in --arg id \"$chat_id\" '.[$id]=true' \"$WF\";fi;local sys=\"\";[ -s \"$SF\" ]&&sys=\" sys:loaded\";echo -e \"\\033[1;30mid:$chat_id msg:$msg_n ts:$(date -u +\"%Y%m%d_%H%M%Sutc\") toks:$((${h_chars:-0}/4))+$(( ${#p}/4 )) md:$M$sys\\033[0m\">&2;_jq_in --arg t \"$p\" '.+[{\"role\":\"user\",\"parts\":[{\"text\":$t}]}]' \"$HF\"||{ echo \"Resetting corrupt history...\">&2;echo \"[]\">\"$HF\";_jq_in --arg t \"$p\" '.+[{\"role\":\"user\",\"parts\":[{\"text\":$t}]}]' \"$HF\"||return 1;};jq -n --slurpfile h \"$HF\" --arg s \"$(<\"$SF\")\" '{contents:$h[0]}+(if($s|length>0)then{system_instruction:{parts:[{text:$s}]}}else{}end)'>\"$PF\";local ft=\"\" us=\"0\" c_args=(-s --connect-timeout 10 -H \"x-goog-api-key: $k\" -H \"Content-Type: application/json\" -d \"@$PF\");[[ \"$(<\"$STF\" 2>/dev/null)\" == \"1\" ]]&&us=\"1\";[ \"$stream_mode\" -eq 1 ]&&us=1;echo -e \"\\033[1;32m🤖 Gemini ($chat_id):\\033[0m\">&2;if [ \"$us\" == \"1\" ];then >\"$RO\";ERR_TMP=$(mktemp \"$D/err.XXXXXX.tmp\");curl \"${c_args[@]}\" -N --max-time 120 -X POST \"$API/$M:streamGenerateContent?alt=sse\"|jq --unbuffered -R -j 'if startswith(\"data: \")then .[6:]|fromjson?|if .error.message then \"\\(.error.message)\\n\"|halt_error(1) else .candidates[0].content.parts[0].text//empty end elif startswith(\"{\")then fromjson?|.error.message|select(.!=null)|\"\\(.)\\n\"|halt_error(1) else empty end' 2>\"$ERR_TMP\"|tee \"$RO\";echo \"\">&2;if [ -s \"$ERR_TMP\" ];then handle_err \"$(<\"$ERR_TMP\")\";rm -f \"$ERR_TMP\";ERR_TMP=\"\";return 1;fi;rm -f \"$ERR_TMP\";ERR_TMP=\"\";ft=$(<\"$RO\");if [ -z \"$ft\" ];then echo -e \"\\033[1;31m❌ Network Error / Timeout.\\033[0m\">&2;_rollback_last;return 1;fi;else local r;r=$(curl \"${c_args[@]}\" --max-time 30 -X POST \"$API/$M:generateContent\");if [ -z \"$r\" ];then echo -e \"\\033[1;31m❌ Network Error / Timeout.\\033[0m\">&2;_rollback_last;return 1;fi;local er;er=$(jq -r '.error.message//empty'<<<\"$r\");if [ -n \"$er\" ];then handle_err \"$er\";return 1;fi;ft=$(jq -r '.candidates[0].content.parts[0].text//empty'<<<\"$r\");printf \"%s\\n\" \"$ft\";if [ -z \"$ft\" ];then echo -e \"\\033[1;33m⚠️ Empty response (safety filter/quota).\\033[0m\">&2;_rollback_last;return 1;fi;fi;[ -n \"$ft\" ]&&_jq_in --arg t \"$ft\" '.+[{\"role\":\"model\",\"parts\":[{\"text\":$t}]}]' \"$HF\";if [ \"$continuous\" -eq 0 ]&&[ ${#MACRO_Q[@]} -eq 0 ];then break;fi;p=\"\";done;}\n[[ \"${BASH_SOURCE[0]}\" == \"${0}\" ]]&&askai \"$@\"\n" } ] }, { "name": "text", "type": "folder", "size": "1 bytes", "path": "text", "children": [ { "name": "foldermake.txt", "type": "file", "size": "1 bytes", "path": "text/foldermake.txt", "content": "\n" } ] }, { "name": "things", "type": "folder", "size": "6,337 bytes", "path": "things", "children": [ { "name": "esp32", "type": "folder", "size": "6,337 bytes", "path": "things/esp32", "children": [ { "name": "sentry1.cpp", "type": "file", "size": "6,337 bytes", "path": "things/esp32/sentry1.cpp", "content": "#include \n#include \n#include \n#include \n#define R return\n#define C esp_wifi_set_\n#define Z(v) if(v>3.14159)v-=6.28318;if(v<-3.14159)v+=6.28318\nstruct V{float x,p,a,o,lp,ld;int s,zc;};\nstruct L{uint32_t t;float p,v;};\nstruct F{float mn,mx;int pt;};\nV m[64];L h[200];int hx=0;uint32_t T;bool Ld=0,Cmp=0;\nWebServer S(80);int i;int8_t b[512];float l,d,cm,tm,iv,Q,al,av;F flt={0.5,50.0,5};\nuint8_t pkt[]={0xc0,0x00,0x00,0x00,0xff,0xff,0xff,0xff,0xff,0xff,0x00,0x00,0x00,0x00,0x00,0x00,0xff,0xff,0xff,0xff,0xff,0xff,0x00,0x00};\nconst char* P=R\"===(\n\n
    \n

    SENTINEL RADAR

    \n
    \n STATUS: INIT\n LIVE MASS: 0.0\n
    \n
    \n \n
    \n
    \n FILTERS (0 = OFF)\n
    \n MIN SIZE:\n
    \n
    \n MAX SIZE:\n
    \n
    \n FAN/OSC:\n
    \n \n
    \n \n
    \n)===\";\nvoid r(void*_,wifi_csi_info_t*o){if(o->len>256)R;memcpy(b,o->buf,o->len);l=0;cm=0;tm=0;\nfor(i=0;i<64;i++){float raw=atan2(b[i*2+1],b[i*2]);if(i){d=raw-l;Z(d);raw=m[i-1].x+d;}l=raw;\nfloat delta=raw-m[i].lp;if((delta>0&&m[i].ld<0)||(delta<0&&m[i].ld>0))m[i].zc++;\nif(m[i].zc>0&&rand()%10==0)m[i].zc--;m[i].lp=raw;m[i].ld=delta;\niv=raw-m[i].x;Q=iv*iv;if(Q>.5)Q=1.;else if(Q<.001)Q=1e-6;\nm[i].p+=Q;float k=m[i].p/(m[i].p+.1);m[i].x+=k*iv;m[i].p=(1.-k)*m[i].p;\nif(!Ld){m[i].a=m[i].x;m[i].o=m[i].x;m[i].s=0;}else{\nif(flt.pt>0&&m[i].zc>flt.pt){m[i].s=3;continue;}\nfloat rf=Cmp?m[i].o:m[i].a;float df=fabs(m[i].x-rf);\nif(df>.5){m[i].s=(m[i].p<.01)?1:2;if(m[i].s==2){cm+=i*df;tm+=df;}}\nelse{m[i].s=0;if(!Cmp)m[i].a=m[i].a*.999+m[i].x*.001;}}}\nav=0;if(Ld){if(flt.mn>0&&tm0&&tm>flt.mx)tm=0;\nif(tm>0){float p=cm/tm,v=0;int lx=hx?hx-1:199;al=p;av=1;\nif(h[lx].t>0){float dt=(millis()-h[lx].t)/1000.;if(dt>0)v=(p-h[lx].p)/dt;}\nh[hx]={millis(),p,v};hx=(hx+1)%200;}}\nvoid setup(){T=millis();WiFi.softAP(\"ESP32_SENTINEL\",\"\");\nfor(i=0;i<64;i++)m[i]={0,1,0,0,0,0,0,0};C csi_rx_cb(r,0);C csi(1);\nwifi_csi_config_t c={1,1,1,1,0,0,0};C csi_config(&c);S.on(\"/\",[](){S.send(200,\"text/html\",P);});\nS.on(\"/s\",[](){if(S.hasArg(\"m\"))Cmp=S.arg(\"m\").toInt();S.send(200,\"text/plain\",\"OK\");});\nS.on(\"/f\",[](){if(S.hasArg(\"n\"))flt.mn=S.arg(\"n\").toFloat();if(S.hasArg(\"x\"))flt.mx=S.arg(\"x\").toFloat();\nif(S.hasArg(\"p\"))flt.pt=S.arg(\"p\").toInt();S.send(200,\"text/plain\",\"OK\");});\nS.on(\"/d\",[](){String j=\"{\\\"s\\\":\\\"\"+String(Ld?\"ARMED\":\"LEARNING\")+\"\\\",\\\"sz\\\":\"+String(tm)+\",\\\"a\\\":\"+String(av)+\",\\\"l\\\":\"+String(al)+\",\\\"m\\\":[\";\nfor(i=0;i<64;i++)j+=(i?\",\":\"\")+String(\"{\\\"x\\\":\")+m[i].x+\",\\\"p\\\":\"+m[i].p+\",\\\"s\\\":\"+m[i].s+\"}\";j+=\"],\\\"h\\\":[\";\nint x=hx;for(int k=0;k<200;k++){if(h[x].t)j+=(k?\",\":\"\")+String(\"{\\\"t\\\":\")+h[x].t+\",\\\"p\\\":\"+h[x].p+\",\\\"v\\\":\"+h[x].v+\"}\";\nx=(x+1)%200;}S.send(200,\"application/json\",j+\"]}\");});S.begin();}\nvoid loop(){S.handleClient();if(!Ld&&millis()-T>60000){Ld=1;for(i=0;i<64;i++)m[i].o=m[i].x;}\nif(Ld){pkt[22]++;esp_wifi_80211_tx(WIFI_IF_AP,pkt,24,1);}delay(5);}\n" } ] } ] }, { "name": "91_character_black_icon.html", "type": "file", "size": "92 bytes", "path": "91_character_black_icon.html", "content": "\n" }, { "name": "enhanced_qr_scanner.html", "type": "file", "size": "62,047 bytes", "path": "enhanced_qr_scanner.html", "content": "\n\n\n \n \n QR Scanner\n \n\n\n\n
    \n
    \n

    QR Scanner

    \n
    \n \n \n \n
    \n
    \n \n \n
    \n Loading...\n 10 Hz\n \n \n \n
    \n
    \n \n
    \n \n 📷 Camera is Off\n
    \n
    \n\n \n
    \n
    \n \n \n \n
    \n \n
    \n
    \n \n Bonus Narrow Focus Region Size: %50\n \n \n
    \n \n
    \n
    \n\n \n
    \n camera mode:\n
    \n \n \n \n \n
    \n
    \n\n
    — OR —
    \n\n \n \n
    \n
    \n\n \n
    \n
    \n

    Scan History

    \n
    \n \n \n \n \n
    RAM: 0 MB
    \n
    \n
    \n
    \n
    \n\n \n
    \n \n \n
    \n \n \n
    \n
    \n

    Editable Links Text

    \n \n
    \n \n \n
    \n
    \n
    \n\n \n
    \n
    \n

    Data Manager (Import / Export)

    \n
    \n

    📤 Download Backup

    \n
    \n \n \n
    \n
    \n \n \n
    \n \n
    \n
    \n

    📥 Upload Backup

    \n
    \n ⚠️ Note: Before using Merge, getting a backup is suggested. It can be aggressive in replacing duplicates.\n
    \n
    \n \n \n
    \n
    \n \n \n
    \n \n
    \n
    \n \n
    \n
    \n
    \n\n \n
    \n
    \n

    seems like you applied BACK

    \n

    \n if it's a mistake click ignore

    \n if you wanted to leave click leave page and use BACK once more\n

    \n
    \n \n \n
    \n

    \n ( if phone stuck you can use BACK 3 times more to return original device behavior which also leaves )\n

    \n
    \n
    \n\n \n \n\n \n\n\n" }, { "name": "enhanced_timer.html", "type": "file", "size": "10,008 bytes", "path": "enhanced_timer.html", "content": "Enhanced Timer
    ms
    00.000
    \n" }, { "name": "file_renamer_and_editor.html", "type": "file", "size": "5,344 bytes", "path": "file_renamer_and_editor.html", "content": "File Renamer and Editor

    Advanced File Editor & Renamer

    Original: No file active

    New: N/A

    File Editor

    ×
    \n" }, { "name": "full_account_via_githack.html", "type": "file", "size": "1,318 bytes", "path": "full_account_via_githack.html", "content": "\n" }, { "name": "headphone_test.html", "type": "file", "size": "3,530 bytes", "path": "headphone_test.html", "content": "Full Earphone Test\n" }, { "name": "HTML_realizer.html", "type": "file", "size": "564 bytes", "path": "HTML_realizer.html", "content": "HTML Realizer
    \n" }, { "name": "html_sites_only_full_account_via_hithack.html", "type": "file", "size": "1,350 bytes", "path": "html_sites_only_full_account_via_hithack.html", "content": "\n" }, { "name": "icon_tester.html", "type": "file", "size": "1,174 bytes", "path": "icon_tester.html", "content": "Icon Tester

    Icon Tester

    Paste a link tag to test a persistent icon.

    \n" }, { "name": "index.html", "type": "file", "size": "4,805 bytes", "path": "index.html", "content": "\n" }, { "name": "primitive_file_comparer.html", "type": "file", "size": "13,697 bytes", "path": "primitive_file_comparer.html", "content": "\n\n\n \n \n Advanced Line Relationship Analyzer\n \n\n\n\n
    \n

    Full Relationship Line Analyzer

    \n
    \n \n \n \n \n
    \n\n \n
    \n

    Overlap Matrix (Who matches Who?)

    \n

    Read this as: \"Rows contain X% of Columns\". e.g., If Row 1 / Col 2 says \"100%\", it means File 1 contains ALL lines from File 2.

    \n
    \n
    \n
    \n
    \n\n \n
    \n

    Common Line \"Buckets\"

    \n

    Click a button to see line math for lines that appear in exactly that many files.

    \n \n
    \n\n
    \n \n
    \n \n

    Sample Lines (First 5 matches in this bucket)

    \n \n
    \n
    \n\n\n\n\n\n" }, { "name": "primitive_file_list_analyser.html", "type": "file", "size": "17,183 bytes", "path": "primitive_file_list_analyser.html", "content": "File list analyser
    Idle
    FILTERS
    Time
    App
    Ext
    Regex
    Slice
    JS
    Exact
    Range
    Cmp
    Path
    Ep/JS
    Prefix
    Interval
    -
    Components (Range ok)
    Path (+)
    Epoch/JS
    App Name
    Extension
    Regex
    Base Sort & Slice
    Custom JS
    ACTIVE
    V: 0T: 0
    TimeAppFilename
    \n" }, { "name": "primitive_gemini_api_chat.html", "type": "file", "size": "13,468 bytes", "path": "primitive_gemini_api_chat.html", "content": "AI Chat
    ms:
    \n" }, { "name": "README.md", "type": "file", "size": "39 bytes", "path": "README.md", "content": "# test11\nversatile sites for laziness \n" }, { "name": "repo_to_json.html", "type": "file", "size": "17,658 bytes", "path": "repo_to_json.html", "content": "Repo to JSON/txt

    Repo to JSON/txt

    Instructions: Configure parameters or use URL.

    Status: Waiting...
    \n" }, { "name": "Speechtotext.html", "type": "file", "size": "3,174 bytes", "path": "Speechtotext.html", "content": "Speech to Text

    Speech-to-Text Tool

    Status: Idle. Click button to grant mic access.

    Easy Copy & Paste Text

    \n" }, { "name": "website_changer.html", "type": "file", "size": "36,883 bytes", "path": "website_changer.html", "content": "Website Changer

    WEBSITE CHANGER

    with
    Proxy: AUTO
    AUTO(loops list in order)
    \n" } ]