Spaces:
Runtime error
Runtime error
Commit
·
a4bb19d
1
Parent(s):
ec2933c
task: init scraper
Browse files- poetry.lock +0 -0
- pyproject.toml +5 -14
- scraper.py +112 -0
- test_scraper.ipynb +103 -0
poetry.lock
ADDED
The diff for this file is too large to render.
See raw diff
|
|
pyproject.toml
CHANGED
@@ -13,6 +13,7 @@ dependencies = [
|
|
13 |
"beautifulsoup4>=4.12.0",
|
14 |
"scrapy>=2.11.0",
|
15 |
"selenium>=4.18.0",
|
|
|
16 |
"playwright>=1.42.0",
|
17 |
"pandas>=2.2.0",
|
18 |
"numpy>=1.26.0",
|
@@ -28,15 +29,9 @@ dependencies = [
|
|
28 |
]
|
29 |
requires-python = ">=3.11,<3.12"
|
30 |
|
31 |
-
[
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
[tool.hatch.build.targets.wheel]
|
36 |
-
packages = ["src/build"]
|
37 |
-
|
38 |
-
[tool.hatch.metadata]
|
39 |
-
allow-direct-references = true
|
40 |
|
41 |
[tool.ruff]
|
42 |
line-length = 88
|
@@ -60,8 +55,4 @@ testpaths = ["tests"]
|
|
60 |
python_version = "3.11"
|
61 |
warn_return_any = true
|
62 |
warn_unused_configs = true
|
63 |
-
check_untyped_defs = true
|
64 |
-
|
65 |
-
[tool.black]
|
66 |
-
line-length = 88
|
67 |
-
target-version = ['py311']
|
|
|
13 |
"beautifulsoup4>=4.12.0",
|
14 |
"scrapy>=2.11.0",
|
15 |
"selenium>=4.18.0",
|
16 |
+
"selenium-wire>=5.1.0", # Added for scraper
|
17 |
"playwright>=1.42.0",
|
18 |
"pandas>=2.2.0",
|
19 |
"numpy>=1.26.0",
|
|
|
29 |
]
|
30 |
requires-python = ">=3.11,<3.12"
|
31 |
|
32 |
+
[tool.black]
|
33 |
+
line-length = 88
|
34 |
+
target-version = ['py311']
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
|
36 |
[tool.ruff]
|
37 |
line-length = 88
|
|
|
55 |
python_version = "3.11"
|
56 |
warn_return_any = true
|
57 |
warn_unused_configs = true
|
58 |
+
check_untyped_defs = true
|
|
|
|
|
|
|
|
scraper.py
ADDED
@@ -0,0 +1,112 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import os
|
2 |
+
import requests
|
3 |
+
from bs4 import BeautifulSoup
|
4 |
+
import json
|
5 |
+
from selenium import webdriver
|
6 |
+
from selenium.webdriver.chrome.options import Options
|
7 |
+
|
8 |
+
def create_design_directory(design_id):
|
9 |
+
"""Create a directory for the design if it doesn't exist"""
|
10 |
+
directory = f"designs/{design_id}"
|
11 |
+
if not os.path.exists(directory):
|
12 |
+
os.makedirs(directory)
|
13 |
+
return directory
|
14 |
+
|
15 |
+
def save_css(url, directory):
|
16 |
+
"""Download and save CSS file"""
|
17 |
+
response = requests.get(url)
|
18 |
+
css_path = f"{directory}/style.css"
|
19 |
+
with open(css_path, "w", encoding="utf-8") as f:
|
20 |
+
f.write(response.text)
|
21 |
+
|
22 |
+
def save_metadata(metadata, directory):
|
23 |
+
"""Save design metadata as JSON"""
|
24 |
+
metadata_path = f"{directory}/metadata.json"
|
25 |
+
with open(metadata_path, "w", encoding="utf-8") as f:
|
26 |
+
json.dump(metadata, f, indent=4)
|
27 |
+
|
28 |
+
def take_screenshot(url, directory):
|
29 |
+
"""Take screenshots of the design at desktop and mobile widths"""
|
30 |
+
chrome_options = Options()
|
31 |
+
chrome_options.add_argument("--headless")
|
32 |
+
|
33 |
+
driver = webdriver.Chrome(options=chrome_options)
|
34 |
+
|
35 |
+
# Desktop screenshot (1920px width)
|
36 |
+
driver.set_window_size(1920, 1080)
|
37 |
+
driver.get(url)
|
38 |
+
# Wait for page to load and get full height
|
39 |
+
total_height = driver.execute_script("return document.body.scrollHeight")
|
40 |
+
driver.set_window_size(1920, total_height)
|
41 |
+
driver.save_screenshot(f"{directory}/screenshot_desktop.png")
|
42 |
+
|
43 |
+
# Mobile screenshot (480px width)
|
44 |
+
driver.set_window_size(480, 1080)
|
45 |
+
driver.get(url)
|
46 |
+
# Wait for page to load and get full height
|
47 |
+
total_height = driver.execute_script("return document.body.scrollHeight")
|
48 |
+
driver.set_window_size(480, total_height)
|
49 |
+
driver.save_screenshot(f"{directory}/screenshot_mobile.png")
|
50 |
+
|
51 |
+
driver.quit()
|
52 |
+
|
53 |
+
def scrape_design(design_id):
|
54 |
+
"""Scrape a single design"""
|
55 |
+
# Create base URLs
|
56 |
+
design_url = f"https://www.csszengarden.com/{design_id}"
|
57 |
+
css_url = f"https://www.csszengarden.com/{design_id}/{design_id}.css"
|
58 |
+
|
59 |
+
# Create directory for this design
|
60 |
+
directory = create_design_directory(design_id)
|
61 |
+
|
62 |
+
# Get design page
|
63 |
+
response = requests.get(design_url)
|
64 |
+
print(f"Response status: {response.status_code}")
|
65 |
+
|
66 |
+
# Debug HTML content
|
67 |
+
print("\nFirst 500 characters of response:")
|
68 |
+
print(response.text[:500])
|
69 |
+
|
70 |
+
soup = BeautifulSoup(response.text, "html.parser")
|
71 |
+
|
72 |
+
# Debug found elements
|
73 |
+
print("\nFound elements:")
|
74 |
+
print(f"h1: {soup.select_one('h1')}")
|
75 |
+
print(f"author: {soup.select_one('meta[name=\"author\"]')}")
|
76 |
+
|
77 |
+
# Extract metadata with error handling
|
78 |
+
try:
|
79 |
+
metadata = {
|
80 |
+
"id": design_id,
|
81 |
+
"author": soup.select_one('meta[name="author"]')["content"] if soup.select_one('meta[name="author"]') else "Unknown Author",
|
82 |
+
"url": design_url,
|
83 |
+
"css_url": css_url
|
84 |
+
}
|
85 |
+
except Exception as e:
|
86 |
+
print(f"\nError extracting metadata: {str(e)}")
|
87 |
+
raise
|
88 |
+
|
89 |
+
# Save everything
|
90 |
+
save_css(css_url, directory)
|
91 |
+
save_metadata(metadata, directory)
|
92 |
+
take_screenshot(design_url, directory)
|
93 |
+
|
94 |
+
def main():
|
95 |
+
"""Main function to scrape multiple designs"""
|
96 |
+
# Create designs directory if it doesn't exist
|
97 |
+
if not os.path.exists("designs"):
|
98 |
+
os.makedirs("designs")
|
99 |
+
|
100 |
+
# List of design IDs to scrape
|
101 |
+
design_ids = ["221", "220", "219"] # Add more IDs as needed
|
102 |
+
|
103 |
+
for design_id in design_ids:
|
104 |
+
try:
|
105 |
+
print(f"Scraping design {design_id}...")
|
106 |
+
scrape_design(design_id)
|
107 |
+
print(f"Successfully scraped design {design_id}")
|
108 |
+
except Exception as e:
|
109 |
+
print(f"Error scraping design {design_id}: {str(e)}")
|
110 |
+
|
111 |
+
if __name__ == "__main__":
|
112 |
+
main()
|
test_scraper.ipynb
ADDED
@@ -0,0 +1,103 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cells": [
|
3 |
+
{
|
4 |
+
"cell_type": "markdown",
|
5 |
+
"metadata": {},
|
6 |
+
"source": [
|
7 |
+
"# Test CSS Zen Garden Scraper\n",
|
8 |
+
"\n",
|
9 |
+
"This notebook tests the functionality of our CSS Zen Garden scraper."
|
10 |
+
]
|
11 |
+
},
|
12 |
+
{
|
13 |
+
"cell_type": "code",
|
14 |
+
"execution_count": 2,
|
15 |
+
"metadata": {},
|
16 |
+
"outputs": [
|
17 |
+
{
|
18 |
+
"name": "stdout",
|
19 |
+
"output_type": "stream",
|
20 |
+
"text": [
|
21 |
+
"Testing scraper with design 221...\n",
|
22 |
+
"Response status: 200\n",
|
23 |
+
"\n",
|
24 |
+
"First 500 characters of response:\n",
|
25 |
+
"<!DOCTYPE html>\n",
|
26 |
+
"<html lang=\"en\">\n",
|
27 |
+
"<head>\n",
|
28 |
+
"\t<meta charset=\"utf-8\">\n",
|
29 |
+
"\t<title>CSS Zen Garden: The Beauty of CSS Design</title>\n",
|
30 |
+
"\n",
|
31 |
+
"\t<link rel=\"stylesheet\" media=\"screen\" href=\"/221/221.css?v=8may2013\">\n",
|
32 |
+
"\t<link rel=\"alternate\" type=\"application/rss+xml\" title=\"RSS\" href=\"http://www.csszengarden.com/zengarden.xml\">\n",
|
33 |
+
"\n",
|
34 |
+
"\t<meta name=\"viewport\" content=\"width=device-width, initial-scale=1.0\">\n",
|
35 |
+
"\t<meta name=\"author\" content=\"Dave Shea\">\n",
|
36 |
+
"\t<meta name=\"description\" content=\"A demonstration of what can be accomplished v\n",
|
37 |
+
"\n",
|
38 |
+
"Found elements:\n",
|
39 |
+
"h1: <h1>CSS Zen Garden</h1>\n",
|
40 |
+
"author: None\n",
|
41 |
+
"author link: None\n",
|
42 |
+
"Success!\n"
|
43 |
+
]
|
44 |
+
}
|
45 |
+
],
|
46 |
+
"source": [
|
47 |
+
"from scraper import create_design_directory, save_css, save_metadata, take_screenshot, scrape_design\n",
|
48 |
+
"\n",
|
49 |
+
"# Test with a single design first\n",
|
50 |
+
"test_design_id = \"221\"\n",
|
51 |
+
"\n",
|
52 |
+
"try:\n",
|
53 |
+
" print(f\"Testing scraper with design {test_design_id}...\")\n",
|
54 |
+
" scrape_design(test_design_id)\n",
|
55 |
+
" print(\"Success!\")\n",
|
56 |
+
"except Exception as e:\n",
|
57 |
+
" print(f\"Error: {str(e)}\")"
|
58 |
+
]
|
59 |
+
},
|
60 |
+
{
|
61 |
+
"cell_type": "code",
|
62 |
+
"execution_count": null,
|
63 |
+
"metadata": {},
|
64 |
+
"outputs": [],
|
65 |
+
"source": [
|
66 |
+
"# If successful, let's check what we got\n",
|
67 |
+
"import json\n",
|
68 |
+
"import os\n",
|
69 |
+
"\n",
|
70 |
+
"design_dir = f\"designs/{test_design_id}\"\n",
|
71 |
+
"print(\"Files created:\")\n",
|
72 |
+
"print(os.listdir(design_dir))\n",
|
73 |
+
"\n",
|
74 |
+
"# Display metadata\n",
|
75 |
+
"with open(f\"{design_dir}/metadata.json\") as f:\n",
|
76 |
+
" metadata = json.load(f)\n",
|
77 |
+
"print(\"\\nMetadata:\")\n",
|
78 |
+
"print(json.dumps(metadata, indent=2))"
|
79 |
+
]
|
80 |
+
}
|
81 |
+
],
|
82 |
+
"metadata": {
|
83 |
+
"kernelspec": {
|
84 |
+
"display_name": ".venv",
|
85 |
+
"language": "python",
|
86 |
+
"name": "python3"
|
87 |
+
},
|
88 |
+
"language_info": {
|
89 |
+
"codemirror_mode": {
|
90 |
+
"name": "ipython",
|
91 |
+
"version": 3
|
92 |
+
},
|
93 |
+
"file_extension": ".py",
|
94 |
+
"mimetype": "text/x-python",
|
95 |
+
"name": "python",
|
96 |
+
"nbconvert_exporter": "python",
|
97 |
+
"pygments_lexer": "ipython3",
|
98 |
+
"version": "3.11.11"
|
99 |
+
}
|
100 |
+
},
|
101 |
+
"nbformat": 4,
|
102 |
+
"nbformat_minor": 2
|
103 |
+
}
|