| import os |
| import re |
|
|
| import pandas as pd |
| from bs4 import BeautifulSoup |
|
|
|
|
| def extract_puzzle_data(html): |
| """extract data from HTML content""" |
| soup = BeautifulSoup(html, 'html.parser') |
|
|
| |
| task_description_section = soup.find('div', id='tabs-2') or "" |
| task_description = task_description_section.get_text(strip=True).replace('\xa0', ' ') |
| task_description = task_description.removeprefix("Backstory and Goal") |
| task_description = task_description.partition("Remember, as with all")[0] |
|
|
| categories = [cc.get_text(strip=True).replace('\xa0', ' ') for cc in soup.find_all('td', class_='answergrid_head')] |
|
|
| |
| clues = [] |
| for clue_div in soup.find_all('div', class_='clue'): |
| clue_raw = clue_div.get_text(strip=True).replace('\xa0', ' ') |
| |
| cleaned_clue = re.sub(r'^\d+\.\s*', '', clue_raw) |
| clues.append(cleaned_clue) |
|
|
| |
| label_categories = dict(label_a=[]) |
| for label in soup.find_all('td', class_='labelboxh'): |
| if label['id'].startswith("labelleftA"): |
| label_categories["label_a"].append(label.get_text(strip=True).replace('\xa0', ' ')) |
| for letter in "bcd": |
| pattern = re.compile(f'label{letter}_ary' + r'\[\d+]\s*=\s*"([^"]+)";') |
| items = pattern.findall(html) |
| label_categories[f"label_{letter}"] = items |
| return dict(story=task_description, |
| clues=clues, |
| categories=categories, |
| **label_categories) |
|
|
|
|
| global_stories = set() |
| global_clues = set() |
|
|
|
|
| def process_one(difficulty, grid_size): |
| puzzle_data = [] |
| with open(f'urls/{difficulty}{grid_size}.txt') as rr: |
| all_paths = [p.strip() for p in rr] |
| dir_path = f'htmls/{difficulty}{grid_size}/' |
| for c, puzzle_url in enumerate(all_paths): |
| filename = puzzle_url.removeprefix("https://logic.puzzlebaron.com/") |
| if c % 200 == 0: |
| print(f"{c=}") |
| file_path = os.path.join(dir_path, filename) |
| with open(file_path, 'r', encoding='utf-8') as file: |
| html_content = file.read() |
| data = extract_puzzle_data(html_content) |
| if len(global_clues.intersection(data['clues'])) >= 3: |
| continue |
| |
| |
| |
| |
| |
| |
| global_clues.update(data['clues']) |
| data['grid_size'] = grid_size |
| data['difficulty'] = difficulty |
| data['url'] = puzzle_url |
| puzzle_data.append(data) |
| return puzzle_data |
|
|
|
|
| OUTPUT_DIR = "dataframes" |
|
|
|
|
| def main(): |
| if not os.path.exists(OUTPUT_DIR): |
| os.makedirs(OUTPUT_DIR) |
| for grid_size in ['4x7', '4x6', '4x5', '4x4', '3x5', '3x4']: |
| for difficulty in ['challenging', 'moderate', 'easy']: |
| puzzle_data = process_one(difficulty, grid_size) |
| df = pd.DataFrame(puzzle_data) |
| jsonl_file_path = f'{OUTPUT_DIR}/{difficulty}{grid_size}.jsonl' |
| df.to_json(jsonl_file_path, orient='records', lines=True) |
| print(f'Data saved to {jsonl_file_path}', df.shape) |
|
|
|
|
| main() |
|
|