3 Commits

Author SHA1 Message Date
calic
a11bdb7267 Stealth mode: defeat Cloudflare automation detection
Some checks failed
Release / build (push) Failing after 1m5s
Drop DefaultExecAllocatorOptions (includes --enable-automation),
add disable-blink-features=AutomationControlled, patch
navigator.webdriver via JS, and inject sessionKey cookie via CDP
into a dedicated fetch profile.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-03-21 00:42:03 +01:00
calic
5abdee06ff Add diagnostic logging to Chrome fallback
All checks were successful
Release / build (push) Successful in 1m32s
Log navigation, polling state, and response snippets so we can
diagnose whether the fallback fails due to Cloudflare challenge,
login redirect, profile lock, or other issues.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-03-21 00:34:41 +01:00
calic
2cb89d3c54 Fix Cloudflare headless detection: use non-headless with hidden window
All checks were successful
Release / build (push) Successful in 1m37s
Cloudflare detects headless Chrome and loops the JS challenge forever.
Switch to non-headless mode with an off-screen window. Also save
Cloudflare cookies (cf_clearance, __cf_bm) after Chrome fallback so
subsequent plain HTTP requests can reuse them.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-03-21 00:26:34 +01:00
2 changed files with 164 additions and 37 deletions

View File

@@ -3,34 +3,54 @@ package browser
import ( import (
"context" "context"
"fmt" "fmt"
"log"
"os" "os"
"path/filepath" "path/filepath"
"strings"
"time" "time"
"github.com/chromedp/cdproto/network"
"github.com/chromedp/chromedp" "github.com/chromedp/chromedp"
"git.davoryn.de/calic/claude-statusline/internal/config" "git.davoryn.de/calic/claude-statusline/internal/config"
) )
// FetchViaChrome navigates to a URL using headless Chrome with the persistent // FetchViaChrome navigates to a URL using Chrome with a dedicated browser
// browser profile (which has Cloudflare clearance cookies) and returns the // profile and returns the response body. Uses stealth flags to avoid
// response body. This bypasses Cloudflare JS challenges because Chrome runs // Cloudflare's automation detection (navigator.webdriver, etc.).
// real JavaScript. Falls back to non-headless if headless fails.
func FetchViaChrome(url string) ([]byte, error) { func FetchViaChrome(url string) ([]byte, error) {
profileDir := filepath.Join(config.ConfigDir(), "browser-profile") // Use a dedicated fetch profile separate from the login profile.
profileDir := filepath.Join(config.ConfigDir(), "fetch-profile")
if err := os.MkdirAll(profileDir, 0o755); err != nil { if err := os.MkdirAll(profileDir, 0o755); err != nil {
return nil, fmt.Errorf("create browser profile dir: %w", err) return nil, fmt.Errorf("create fetch profile dir: %w", err)
} }
// Remove stale lock file from unclean shutdown
_ = os.Remove(filepath.Join(profileDir, "SingletonLock")) _ = os.Remove(filepath.Join(profileDir, "SingletonLock"))
execPath := findBrowserExec() execPath := findBrowserExec()
opts := append(chromedp.DefaultExecAllocatorOptions[:], // Start with minimal options — NOT DefaultExecAllocatorOptions, which
chromedp.Flag("headless", true), // includes flags that Cloudflare detects (--enable-automation, etc.).
opts := []chromedp.ExecAllocatorOption{
chromedp.NoFirstRun,
chromedp.NoDefaultBrowserCheck,
chromedp.UserDataDir(profileDir), chromedp.UserDataDir(profileDir),
)
// Stealth: disable automation indicators
chromedp.Flag("disable-blink-features", "AutomationControlled"),
chromedp.Flag("enable-automation", false),
chromedp.Flag("disable-infobars", true),
// Window: off-screen so it doesn't flash
chromedp.Flag("window-position", "-32000,-32000"),
chromedp.Flag("window-size", "1,1"),
chromedp.Flag("disable-gpu", true),
chromedp.Flag("disable-extensions", true),
chromedp.Flag("no-first-run", true),
// Use non-headless — Cloudflare detects headless mode
chromedp.Flag("headless", false),
}
if execPath != "" { if execPath != "" {
opts = append(opts, chromedp.ExecPath(execPath)) opts = append(opts, chromedp.ExecPath(execPath))
} }
@@ -41,36 +61,112 @@ func FetchViaChrome(url string) ([]byte, error) {
ctx, cancel := chromedp.NewContext(allocCtx) ctx, cancel := chromedp.NewContext(allocCtx)
defer cancel() defer cancel()
// Set a total timeout for the operation
ctx, timeoutCancel := context.WithTimeout(ctx, 30*time.Second) ctx, timeoutCancel := context.WithTimeout(ctx, 30*time.Second)
defer timeoutCancel() defer timeoutCancel()
var body string // Inject the session key cookie before navigating
sessionKey := config.GetSessionKey()
if sessionKey != "" {
err := chromedp.Run(ctx, err := chromedp.Run(ctx,
chromedp.Navigate(url), chromedp.Navigate("about:blank"),
// Wait for the body to have content (Cloudflare challenge resolves via JS) setCookieAction("sessionKey", sessionKey, ".claude.ai"),
chromedp.WaitReady("body"),
// Chrome renders JSON API responses inside a <pre> tag
chromedp.Text("pre", &body, chromedp.ByQuery, chromedp.NodeVisible),
) )
if err != nil { if err != nil {
// Fallback: try extracting from body directly (some responses may not use <pre>) log.Printf("chrome-fetch: cookie injection failed: %v", err)
var bodyFallback string }
errFb := chromedp.Run(ctx, }
chromedp.Text("body", &bodyFallback, chromedp.ByQuery),
// Patch navigator.webdriver to false via CDP
err := chromedp.Run(ctx,
chromedp.ActionFunc(func(ctx context.Context) error {
_, _, _, err := chromedp.Targets(ctx)
_ = err
return nil
}),
chromedp.Evaluate(`Object.defineProperty(navigator, 'webdriver', {get: () => undefined})`, nil),
) )
if errFb == nil && bodyFallback != "" { if err != nil {
return []byte(bodyFallback), nil log.Printf("chrome-fetch: webdriver patch failed: %v", err)
}
return nil, fmt.Errorf("chromedp fetch: %w", err)
} }
log.Printf("chrome-fetch: navigating to %s (exec: %q)", url, execPath)
if err := chromedp.Run(ctx, chromedp.Navigate(url)); err != nil {
log.Printf("chrome-fetch: navigate failed: %v", err)
return nil, fmt.Errorf("chromedp navigate: %w", err)
}
log.Printf("chrome-fetch: navigation complete, polling for JSON...")
// Poll for JSON response — Cloudflare challenge takes a few seconds
ticker := time.NewTicker(1 * time.Second)
defer ticker.Stop()
for {
select {
case <-ctx.Done():
return nil, fmt.Errorf("chromedp fetch timed out waiting for JSON response")
case <-ticker.C:
var body string
err := chromedp.Run(ctx,
chromedp.Text("pre", &body, chromedp.ByQuery),
)
if err != nil || body == "" {
_ = chromedp.Run(ctx,
chromedp.Text("body", &body, chromedp.ByQuery),
)
}
body = strings.TrimSpace(body)
if body == "" { if body == "" {
return nil, fmt.Errorf("chromedp fetch: empty response body") log.Printf("chrome-fetch: page body empty, waiting...")
continue
}
if body[0] == '[' || body[0] == '{' {
log.Printf("chrome-fetch: got JSON response (%d bytes)", len(body))
_ = extractAndSaveCookies(ctx)
cancel()
return []byte(body), nil
}
snippet := body
if len(snippet) > 200 {
snippet = snippet[:200]
}
log.Printf("chrome-fetch: non-JSON body (%d bytes): %s", len(body), snippet)
}
}
}
// setCookieAction sets a cookie via the DevTools protocol.
func setCookieAction(name, value, domain string) chromedp.Action {
return chromedp.ActionFunc(func(ctx context.Context) error {
expr := network.SetCookie(name, value).
WithDomain(domain).
WithPath("/").
WithHTTPOnly(true).
WithSecure(true)
return expr.Do(ctx)
})
}
// extractAndSaveCookies saves cf_clearance and other Cloudflare cookies
// so plain HTTP requests can try them on subsequent polls.
func extractAndSaveCookies(ctx context.Context) error {
cookies, err := network.GetCookies().Do(ctx)
if err != nil {
return err
} }
// Gracefully close to flush cookies (including refreshed cf_clearance) var parts []string
cancel() for _, c := range cookies {
if c.Domain == ".claude.ai" || c.Domain == "claude.ai" {
if c.Name == "cf_clearance" || c.Name == "__cf_bm" || c.Name == "_cfuvid" {
parts = append(parts, c.Name+"="+c.Value)
}
}
}
return []byte(body), nil if len(parts) == 0 {
return nil
}
cfPath := filepath.Join(config.ConfigDir(), "cf-cookies")
return os.WriteFile(cfPath, []byte(strings.Join(parts, "\n")+"\n"), 0o600)
} }

View File

@@ -7,6 +7,9 @@ import (
"log" "log"
"math" "math"
"net/http" "net/http"
"os"
"path/filepath"
"strings"
"sync" "sync"
"time" "time"
@@ -34,13 +37,21 @@ type ParsedUsage struct {
type UpdateCallback func(ParsedUsage) type UpdateCallback func(ParsedUsage)
// doRequest performs an authenticated HTTP GET to the Claude API. // doRequest performs an authenticated HTTP GET to the Claude API.
// Includes any saved Cloudflare cookies from previous Chrome fallbacks.
func doRequest(url, sessionKey string) ([]byte, int, error) { func doRequest(url, sessionKey string) ([]byte, int, error) {
client := &http.Client{Timeout: 10 * time.Second} client := &http.Client{Timeout: 10 * time.Second}
req, err := http.NewRequest("GET", url, nil) req, err := http.NewRequest("GET", url, nil)
if err != nil { if err != nil {
return nil, 0, err return nil, 0, err
} }
req.Header.Set("Cookie", "sessionKey="+sessionKey)
cookie := "sessionKey=" + sessionKey
// Append Cloudflare cookies if available (saved by Chrome fallback)
if cfCookies := loadCFCookies(); cfCookies != "" {
cookie += "; " + cfCookies
}
req.Header.Set("Cookie", cookie)
req.Header.Set("User-Agent", userAgent) req.Header.Set("User-Agent", userAgent)
req.Header.Set("Accept", "application/json") req.Header.Set("Accept", "application/json")
req.Header.Set("Referer", "https://claude.ai/") req.Header.Set("Referer", "https://claude.ai/")
@@ -58,6 +69,24 @@ func doRequest(url, sessionKey string) ([]byte, int, error) {
return body, resp.StatusCode, nil return body, resp.StatusCode, nil
} }
// loadCFCookies reads saved Cloudflare cookies from the cf-cookies file.
func loadCFCookies() string {
data, err := os.ReadFile(filepath.Join(config.ConfigDir(), "cf-cookies"))
if err != nil {
return ""
}
// File has one cookie per line (name=value), join with "; "
lines := strings.Split(strings.TrimSpace(string(data)), "\n")
var valid []string
for _, l := range lines {
l = strings.TrimSpace(l)
if l != "" {
valid = append(valid, l)
}
}
return strings.Join(valid, "; ")
}
// fetchWithFallback tries a plain HTTP request first, then falls back to // fetchWithFallback tries a plain HTTP request first, then falls back to
// headless Chrome (which can solve Cloudflare JS challenges) on 403. // headless Chrome (which can solve Cloudflare JS challenges) on 403.
func fetchWithFallback(url, sessionKey string) ([]byte, error) { func fetchWithFallback(url, sessionKey string) ([]byte, error) {
@@ -73,11 +102,13 @@ func fetchWithFallback(url, sessionKey string) ([]byte, error) {
} }
if status == 403 { if status == 403 {
// Likely a Cloudflare JS challenge — fall back to headless Chrome // Likely a Cloudflare JS challenge — fall back to headless Chrome
log.Printf("HTTP 403 for %s, falling back to headless Chrome", url) log.Printf("HTTP 403 for %s, falling back to Chrome", url)
chromeBody, chromeErr := browser.FetchViaChrome(url) chromeBody, chromeErr := browser.FetchViaChrome(url)
if chromeErr != nil { if chromeErr != nil {
return nil, fmt.Errorf("auth_expired") // treat as auth failure if Chrome also fails log.Printf("Chrome fallback failed: %v", chromeErr)
return nil, fmt.Errorf("auth_expired")
} }
log.Printf("Chrome fallback succeeded (%d bytes)", len(chromeBody))
return chromeBody, nil return chromeBody, nil
} }
return nil, fmt.Errorf("HTTP %d", status) return nil, fmt.Errorf("HTTP %d", status)