diff --git a/doc/cp.nvim.txt b/doc/cp.nvim.txt index b4a00d3..07d4312 100644 --- a/doc/cp.nvim.txt +++ b/doc/cp.nvim.txt @@ -28,7 +28,7 @@ COMMANDS *cp-commands* cp.nvim uses a single :CP command with intelligent argument parsing: State Restoration ~ - :CP Restore contest context from current file. + :CP Restore state from current file. Automatically detects platform, contest, problem, and language from cached state. Use this after switching files to restore your CP environment. diff --git a/scrapers/codeforces.py b/scrapers/codeforces.py index d98e657..94abf85 100644 --- a/scrapers/codeforces.py +++ b/scrapers/codeforces.py @@ -7,7 +7,7 @@ from dataclasses import asdict import requests from bs4 import BeautifulSoup, Tag -from scrapling.fetchers import StealthySession +from scrapling.fetchers import StealthyFetcher from .base import BaseScraper from .models import ( @@ -22,9 +22,8 @@ from .models import ( def scrape(url: str) -> list[TestCase]: try: - with StealthySession(headless=True, solve_cloudflare=True) as session: - page = session.fetch(url, google_search=False) - html = page.html_content + page = StealthyFetcher.fetch(url, headless=True, solve_cloudflare=True) + html = page.html_content soup = BeautifulSoup(html, "html.parser") input_sections = soup.find_all("div", class_="input") @@ -181,9 +180,8 @@ def extract_problem_limits(soup: BeautifulSoup) -> tuple[int, float]: def scrape_contest_problems(contest_id: str) -> list[ProblemSummary]: try: contest_url: str = f"https://codeforces.com/contest/{contest_id}" - with StealthySession(headless=True, solve_cloudflare=True) as session: - page = session.fetch(contest_url, google_search=False) - html = page.html_content + page = StealthyFetcher.fetch(contest_url, headless=True, solve_cloudflare=True) + html = page.html_content soup = BeautifulSoup(html, "html.parser") problems: list[ProblemSummary] = [] @@ -276,9 +274,8 @@ class CodeforcesScraper(BaseScraper): url = parse_problem_url(contest_id, problem_letter) tests = scrape_sample_tests(url) - with StealthySession(headless=True, solve_cloudflare=True) as session: - page = session.fetch(url, google_search=False) - html = page.html_content + page = StealthyFetcher.fetch(url, headless=True, solve_cloudflare=True) + html = page.html_content soup = BeautifulSoup(html, "html.parser") timeout_ms, memory_mb = extract_problem_limits(soup) diff --git a/tests/scrapers/test_codeforces.py b/tests/scrapers/test_codeforces.py index e1e555f..6971ed6 100644 --- a/tests/scrapers/test_codeforces.py +++ b/tests/scrapers/test_codeforces.py @@ -4,18 +4,10 @@ from scrapers.codeforces import CodeforcesScraper from scrapers.models import ContestSummary, ProblemSummary -def make_mock_session(html: str): - """Return a mock StealthySession that yields the given HTML.""" - mock_session = Mock() - mock_session.fetch.return_value.html_content = html - mock_session.__enter__ = lambda s: s - mock_session.__exit__ = lambda s, exc_type, exc_val, exc_tb: None - return mock_session - - def test_scrape_success(mocker, mock_codeforces_html): - mock_session = make_mock_session(mock_codeforces_html) - mocker.patch("scrapers.codeforces.StealthySession", return_value=mock_session) + mock_page = Mock() + mock_page.html_content = mock_codeforces_html + mocker.patch("scrapers.codeforces.StealthyFetcher.fetch", return_value=mock_page) scraper = CodeforcesScraper() result = scraper.scrape_problem_tests("1900", "A") @@ -31,8 +23,9 @@ def test_scrape_contest_problems(mocker): A. Problem A B. Problem B """ - mock_session = make_mock_session(html) - mocker.patch("scrapers.codeforces.StealthySession", return_value=mock_session) + mock_page = Mock() + mock_page.html_content = html + mocker.patch("scrapers.codeforces.StealthyFetcher.fetch", return_value=mock_page) scraper = CodeforcesScraper() result = scraper.scrape_contest_metadata("1900") @@ -44,11 +37,10 @@ def test_scrape_contest_problems(mocker): def test_scrape_network_error(mocker): - mock_session = Mock() - mock_session.fetch.side_effect = Exception("Network error") - mock_session.__enter__ = lambda s: s - mock_session.__exit__ = lambda s, exc_type, exc_val, exc_tb: None - mocker.patch("scrapers.codeforces.StealthySession", return_value=mock_session) + mocker.patch( + "scrapers.codeforces.StealthyFetcher.fetch", + side_effect=Exception("Network error"), + ) scraper = CodeforcesScraper() result = scraper.scrape_problem_tests("1900", "A")