diff --git a/doc/cp.nvim.txt b/doc/cp.nvim.txt
index b4a00d3..07d4312 100644
--- a/doc/cp.nvim.txt
+++ b/doc/cp.nvim.txt
@@ -28,7 +28,7 @@ COMMANDS *cp-commands*
cp.nvim uses a single :CP command with intelligent argument parsing:
State Restoration ~
- :CP Restore contest context from current file.
+ :CP Restore state from current file.
Automatically detects platform, contest, problem,
and language from cached state. Use this after
switching files to restore your CP environment.
diff --git a/scrapers/codeforces.py b/scrapers/codeforces.py
index d98e657..94abf85 100644
--- a/scrapers/codeforces.py
+++ b/scrapers/codeforces.py
@@ -7,7 +7,7 @@ from dataclasses import asdict
import requests
from bs4 import BeautifulSoup, Tag
-from scrapling.fetchers import StealthySession
+from scrapling.fetchers import StealthyFetcher
from .base import BaseScraper
from .models import (
@@ -22,9 +22,8 @@ from .models import (
def scrape(url: str) -> list[TestCase]:
try:
- with StealthySession(headless=True, solve_cloudflare=True) as session:
- page = session.fetch(url, google_search=False)
- html = page.html_content
+ page = StealthyFetcher.fetch(url, headless=True, solve_cloudflare=True)
+ html = page.html_content
soup = BeautifulSoup(html, "html.parser")
input_sections = soup.find_all("div", class_="input")
@@ -181,9 +180,8 @@ def extract_problem_limits(soup: BeautifulSoup) -> tuple[int, float]:
def scrape_contest_problems(contest_id: str) -> list[ProblemSummary]:
try:
contest_url: str = f"https://codeforces.com/contest/{contest_id}"
- with StealthySession(headless=True, solve_cloudflare=True) as session:
- page = session.fetch(contest_url, google_search=False)
- html = page.html_content
+ page = StealthyFetcher.fetch(contest_url, headless=True, solve_cloudflare=True)
+ html = page.html_content
soup = BeautifulSoup(html, "html.parser")
problems: list[ProblemSummary] = []
@@ -276,9 +274,8 @@ class CodeforcesScraper(BaseScraper):
url = parse_problem_url(contest_id, problem_letter)
tests = scrape_sample_tests(url)
- with StealthySession(headless=True, solve_cloudflare=True) as session:
- page = session.fetch(url, google_search=False)
- html = page.html_content
+ page = StealthyFetcher.fetch(url, headless=True, solve_cloudflare=True)
+ html = page.html_content
soup = BeautifulSoup(html, "html.parser")
timeout_ms, memory_mb = extract_problem_limits(soup)
diff --git a/tests/scrapers/test_codeforces.py b/tests/scrapers/test_codeforces.py
index e1e555f..6971ed6 100644
--- a/tests/scrapers/test_codeforces.py
+++ b/tests/scrapers/test_codeforces.py
@@ -4,18 +4,10 @@ from scrapers.codeforces import CodeforcesScraper
from scrapers.models import ContestSummary, ProblemSummary
-def make_mock_session(html: str):
- """Return a mock StealthySession that yields the given HTML."""
- mock_session = Mock()
- mock_session.fetch.return_value.html_content = html
- mock_session.__enter__ = lambda s: s
- mock_session.__exit__ = lambda s, exc_type, exc_val, exc_tb: None
- return mock_session
-
-
def test_scrape_success(mocker, mock_codeforces_html):
- mock_session = make_mock_session(mock_codeforces_html)
- mocker.patch("scrapers.codeforces.StealthySession", return_value=mock_session)
+ mock_page = Mock()
+ mock_page.html_content = mock_codeforces_html
+ mocker.patch("scrapers.codeforces.StealthyFetcher.fetch", return_value=mock_page)
scraper = CodeforcesScraper()
result = scraper.scrape_problem_tests("1900", "A")
@@ -31,8 +23,9 @@ def test_scrape_contest_problems(mocker):
A. Problem A
B. Problem B
"""
- mock_session = make_mock_session(html)
- mocker.patch("scrapers.codeforces.StealthySession", return_value=mock_session)
+ mock_page = Mock()
+ mock_page.html_content = html
+ mocker.patch("scrapers.codeforces.StealthyFetcher.fetch", return_value=mock_page)
scraper = CodeforcesScraper()
result = scraper.scrape_contest_metadata("1900")
@@ -44,11 +37,10 @@ def test_scrape_contest_problems(mocker):
def test_scrape_network_error(mocker):
- mock_session = Mock()
- mock_session.fetch.side_effect = Exception("Network error")
- mock_session.__enter__ = lambda s: s
- mock_session.__exit__ = lambda s, exc_type, exc_val, exc_tb: None
- mocker.patch("scrapers.codeforces.StealthySession", return_value=mock_session)
+ mocker.patch(
+ "scrapers.codeforces.StealthyFetcher.fetch",
+ side_effect=Exception("Network error"),
+ )
scraper = CodeforcesScraper()
result = scraper.scrape_problem_tests("1900", "A")