__package__ = 'archivebox.core' import os import sys from django.utils import timezone import inspect from typing import Callable, get_type_hints from pathlib import Path from django.shortcuts import render, redirect from django.http import HttpRequest, HttpResponse, Http404 from django.utils.html import format_html, mark_safe from django.views import View from django.views.generic.list import ListView from django.views.generic import FormView from django.db.models import Q from django.contrib import messages from django.contrib.auth.mixins import UserPassesTestMixin from django.views.decorators.csrf import csrf_exempt from django.utils.decorators import method_decorator from admin_data_views.typing import TableContext, ItemContext from admin_data_views.utils import render_with_table_view, render_with_item_view, ItemLink import archivebox from archivebox.config import CONSTANTS, CONSTANTS_CONFIG, DATA_DIR, VERSION from archivebox.config.common import SHELL_CONFIG, SERVER_CONFIG, ARCHIVING_CONFIG from archivebox.config.configset import get_flat_config, get_config, get_all_configs from archivebox.misc.util import base_url, htmlencode, ts_to_date_str from archivebox.misc.serve_static import serve_static_with_byterange_support from archivebox.misc.logging_util import printable_filesize from archivebox.search import query_search_index from archivebox.core.models import Snapshot from archivebox.core.forms import AddLinkForm from archivebox.crawls.models import Crawl from archivebox.hooks import get_enabled_plugins, get_plugin_name class HomepageView(View): def get(self, request): if request.user.is_authenticated: return redirect('/admin/core/snapshot/') if SERVER_CONFIG.PUBLIC_INDEX: return redirect('/public') return redirect(f'/admin/login/?next={request.path}') class SnapshotView(View): # render static html index from filesystem archive//index.html @staticmethod def render_live_index(request, snapshot): TITLE_LOADING_MSG = 'Not yet archived...' # Dict of plugin -> ArchiveResult object archiveresult_objects = {} # Dict of plugin -> result info dict (for template compatibility) archiveresults = {} results = snapshot.archiveresult_set.all() for result in results: embed_path = result.embed_path() abs_path = result.snapshot_dir / (embed_path or 'None') if (result.status == 'succeeded' and embed_path and os.access(abs_path, os.R_OK) and abs_path.exists()): if os.path.isdir(abs_path) and not any(abs_path.glob('*.*')): continue # Store the full ArchiveResult object for template tags archiveresult_objects[result.plugin] = result result_info = { 'name': result.plugin, 'path': embed_path, 'ts': ts_to_date_str(result.end_ts), 'size': abs_path.stat().st_size or '?', 'result': result, # Include the full object for template tags } archiveresults[result.plugin] = result_info # Use canonical_outputs for intelligent discovery # This method now scans ArchiveResults and uses smart heuristics canonical = snapshot.canonical_outputs() # Add any newly discovered outputs from canonical_outputs to archiveresults snap_dir = Path(snapshot.output_dir) for key, path in canonical.items(): if not key.endswith('_path') or not path or path.startswith('http'): continue plugin_name = key.replace('_path', '') if plugin_name in archiveresults: continue # Already have this from ArchiveResult file_path = snap_dir / path if not file_path.exists() or not file_path.is_file(): continue try: file_size = file_path.stat().st_size if file_size >= 15_000: # Only show files > 15KB archiveresults[plugin_name] = { 'name': plugin_name, 'path': path, 'ts': ts_to_date_str(file_path.stat().st_mtime or 0), 'size': file_size, 'result': None, } except OSError: continue # Get available extractor plugins from hooks (sorted by numeric prefix for ordering) # Convert to base names for display ordering all_plugins = [get_plugin_name(e) for e in get_enabled_plugins()] preferred_types = tuple(all_plugins) all_types = preferred_types + tuple(result_type for result_type in archiveresults.keys() if result_type not in preferred_types) best_result = {'path': 'None', 'result': None} for result_type in preferred_types: if result_type in archiveresults: best_result = archiveresults[result_type] break snapshot_info = snapshot.to_dict(extended=True) try: warc_path = 'warc/' + list(Path(snap_dir).glob('warc/*.warc.*'))[0].name except IndexError: warc_path = 'warc/' context = { **snapshot_info, **snapshot_info.get('canonical', {}), 'title': htmlencode( snapshot.title or (snapshot.base_url if snapshot.is_archived else TITLE_LOADING_MSG) ), 'extension': snapshot.extension or 'html', 'tags': snapshot.tags_str() or 'untagged', 'size': printable_filesize(snapshot.archive_size) if snapshot.archive_size else 'pending', 'status': 'archived' if snapshot.is_archived else 'not yet archived', 'status_color': 'success' if snapshot.is_archived else 'danger', 'oldest_archive_date': ts_to_date_str(snapshot.oldest_archive_date), 'warc_path': warc_path, 'PREVIEW_ORIGINALS': SERVER_CONFIG.PREVIEW_ORIGINALS, 'archiveresults': sorted(archiveresults.values(), key=lambda r: all_types.index(r['name']) if r['name'] in all_types else -r['size']), 'best_result': best_result, 'snapshot': snapshot, # Pass the snapshot object for template tags } return render(template_name='core/snapshot_live.html', request=request, context=context) def get(self, request, path): if not request.user.is_authenticated and not SERVER_CONFIG.PUBLIC_SNAPSHOTS: return redirect(f'/admin/login/?next={request.path}') snapshot = None try: slug, archivefile = path.split('/', 1) except (IndexError, ValueError): slug, archivefile = path.split('/', 1)[0], 'index.html' # slug is a timestamp if slug.replace('.','').isdigit(): # missing trailing slash -> redirect to index if '/' not in path: return redirect(f'{path}/index.html') try: try: snapshot = Snapshot.objects.get(Q(timestamp=slug) | Q(id__startswith=slug)) if archivefile == 'index.html': # if they requested snapshot index, serve live rendered template instead of static html response = self.render_live_index(request, snapshot) else: response = serve_static_with_byterange_support( request, archivefile, document_root=snapshot.output_dir, show_indexes=True, ) response["Link"] = f'<{snapshot.url}>; rel="canonical"' return response except Snapshot.DoesNotExist: if Snapshot.objects.filter(timestamp__startswith=slug).exists(): raise Snapshot.MultipleObjectsReturned else: raise except Snapshot.DoesNotExist: # Snapshot does not exist return HttpResponse( format_html( ( '



' 'No Snapshot directories match the given timestamp/ID: {}

' 'You can add a new Snapshot, or return to the Main Index' '
' ), slug, path, ), content_type="text/html", status=404, ) except Snapshot.MultipleObjectsReturned: snapshot_hrefs = mark_safe('
').join( format_html( '{} {} {} {}', snap.bookmarked_at.strftime('%Y-%m-%d %H:%M:%S'), snap.timestamp, snap.timestamp, snap.url, snap.title_stripped[:64] or '', ) for snap in Snapshot.objects.filter(timestamp__startswith=slug).only('url', 'timestamp', 'title', 'bookmarked_at').order_by('-bookmarked_at') ) return HttpResponse( format_html( ( 'Multiple Snapshots match the given timestamp/ID {}
'
                        ),
                        slug,
                    ) + snapshot_hrefs + format_html(
                        (
                            '

' 'Choose a Snapshot to proceed or go back to the Main Index' ) ), content_type="text/html", status=404, ) except Http404: assert snapshot # (Snapshot.DoesNotExist is already handled above) # Snapshot dir exists but file within does not e.g. 124235.324234/screenshot.png return HttpResponse( format_html( ( '' 'Snapshot Not Found' #'' '' '



' f'Snapshot [{snapshot.timestamp}]: {snapshot.url}
' f'was queued on {str(snapshot.bookmarked_at).split(".")[0]}, ' f'but no files have been saved yet in:
{snapshot.timestamp}/' '{}' f'

' 'It\'s possible {} ' f'during the last capture on {str(snapshot.bookmarked_at).split(".")[0]},
or that the archiving process has not completed yet.
' f'
# run this cmd to finish/retry archiving this Snapshot
' f'archivebox update -t timestamp {snapshot.timestamp}


' '
' 'Next steps:
' f'- list all the Snapshot files .*
' f'- view the Snapshot ./index.html
' f'- go to the Snapshot admin to edit
' f'- go to the Snapshot actions to re-archive
' '- or return to the main index...
' '
' '' ), archivefile if str(archivefile) != 'None' else '', f'the {archivefile} resource could not be fetched' if str(archivefile) != 'None' else 'the original site was not available', ), content_type="text/html", status=404, ) # slug is a URL try: try: # try exact match on full url / ID first snapshot = Snapshot.objects.get( Q(url='http://' + path) | Q(url='https://' + path) | Q(id__icontains=path) ) except Snapshot.DoesNotExist: # fall back to match on exact base_url try: snapshot = Snapshot.objects.get( Q(url='http://' + base_url(path)) | Q(url='https://' + base_url(path)) ) except Snapshot.DoesNotExist: # fall back to matching base_url as prefix snapshot = Snapshot.objects.get( Q(url__startswith='http://' + base_url(path)) | Q(url__startswith='https://' + base_url(path)) ) return redirect(f'/archive/{snapshot.timestamp}/index.html') except Snapshot.DoesNotExist: return HttpResponse( format_html( ( '



' 'No Snapshots match the given url: {}


' 'Return to the Main Index, or:

' '+ Add a new Snapshot for {}

' '
' ), base_url(path), path if '://' in path else f'https://{path}', path, ), content_type="text/html", status=404, ) except Snapshot.MultipleObjectsReturned: snapshot_hrefs = mark_safe('
').join( format_html( '{} {} {} {} {}', snap.bookmarked_at.strftime('%Y-%m-%d %H:%M:%S'), str(snap.id)[:8], snap.timestamp, snap.timestamp, snap.url, snap.title_stripped[:64] or '', ) for snap in Snapshot.objects.filter( Q(url__startswith='http://' + base_url(path)) | Q(url__startswith='https://' + base_url(path)) | Q(id__icontains=path) ).only('url', 'timestamp', 'title', 'bookmarked_at').order_by('-bookmarked_at') ) return HttpResponse( format_html( ( 'Multiple Snapshots match the given URL {}
'
                    ),
                    base_url(path),
                ) + snapshot_hrefs + format_html(
                    (
                        '

' 'Choose a Snapshot to proceed or go back to the Main Index' ) ), content_type="text/html", status=404, ) class PublicIndexView(ListView): template_name = 'public_index.html' model = Snapshot paginate_by = SERVER_CONFIG.SNAPSHOTS_PER_PAGE ordering = ['-bookmarked_at', '-created_at'] def get_context_data(self, **kwargs): return { **super().get_context_data(**kwargs), 'VERSION': VERSION, 'COMMIT_HASH': SHELL_CONFIG.COMMIT_HASH, 'FOOTER_INFO': SERVER_CONFIG.FOOTER_INFO, } def get_queryset(self, **kwargs): qs = super().get_queryset(**kwargs) query = self.request.GET.get('q', default = '').strip() if not query: return qs.distinct() query_type = self.request.GET.get('query_type') if not query_type or query_type == 'all': qs = qs.filter(Q(title__icontains=query) | Q(url__icontains=query) | Q(timestamp__icontains=query) | Q(tags__name__icontains=query)) try: qs = qs | query_search_index(query) except Exception as err: print(f'[!] Error while using search backend: {err.__class__.__name__} {err}') elif query_type == 'fulltext': try: qs = qs | query_search_index(query) except Exception as err: print(f'[!] Error while using search backend: {err.__class__.__name__} {err}') elif query_type == 'meta': qs = qs.filter(Q(title__icontains=query) | Q(url__icontains=query) | Q(timestamp__icontains=query) | Q(tags__name__icontains=query)) elif query_type == 'url': qs = qs.filter(Q(url__icontains=query)) elif query_type == 'title': qs = qs.filter(Q(title__icontains=query)) elif query_type == 'timestamp': qs = qs.filter(Q(timestamp__icontains=query)) elif query_type == 'tags': qs = qs.filter(Q(tags__name__icontains=query)) else: print(f'[!] Unknown value for query_type: "{query_type}"') return qs.distinct() def get(self, *args, **kwargs): if SERVER_CONFIG.PUBLIC_INDEX or self.request.user.is_authenticated: response = super().get(*args, **kwargs) return response else: return redirect(f'/admin/login/?next={self.request.path}') @method_decorator(csrf_exempt, name='dispatch') class AddView(UserPassesTestMixin, FormView): template_name = "add.html" form_class = AddLinkForm def get_initial(self): """Prefill the AddLinkForm with the 'url' GET parameter""" if self.request.method == 'GET': url = self.request.GET.get('url', None) if url: return {'url': url if '://' in url else f'https://{url}'} return super().get_initial() def test_func(self): return SERVER_CONFIG.PUBLIC_ADD_VIEW or self.request.user.is_authenticated def get_context_data(self, **kwargs): from archivebox.core.models import Tag return { **super().get_context_data(**kwargs), 'title': "Create Crawl", # We can't just call request.build_absolute_uri in the template, because it would include query parameters 'absolute_add_path': self.request.build_absolute_uri(self.request.path), 'VERSION': VERSION, 'FOOTER_INFO': SERVER_CONFIG.FOOTER_INFO, 'stdout': '', 'available_tags': list(Tag.objects.all().order_by('name').values_list('name', flat=True)), } def form_valid(self, form): urls = form.cleaned_data["url"] print(f'[+] Adding URL: {urls}') # Extract all form fields tag = form.cleaned_data["tag"] depth = int(form.cleaned_data["depth"]) plugins = ','.join(form.cleaned_data.get("plugins", [])) schedule = form.cleaned_data.get("schedule", "").strip() persona = form.cleaned_data.get("persona", "Default") overwrite = form.cleaned_data.get("overwrite", False) update = form.cleaned_data.get("update", False) index_only = form.cleaned_data.get("index_only", False) notes = form.cleaned_data.get("notes", "") custom_config = form.cleaned_data.get("config", {}) from archivebox.config.permissions import HOSTNAME # 1. save the provided urls to sources/2024-11-05__23-59-59__web_ui_add_by_user_.txt sources_file = CONSTANTS.SOURCES_DIR / f'{timezone.now().strftime("%Y-%m-%d__%H-%M-%S")}__web_ui_add_by_user_{self.request.user.pk}.txt' sources_file.write_text(urls if isinstance(urls, str) else '\n'.join(urls)) # 2. create a new Crawl with the URLs from the file timestamp = timezone.now().strftime("%Y-%m-%d__%H-%M-%S") urls_content = sources_file.read_text() # Build complete config config = { 'ONLY_NEW': not update, 'INDEX_ONLY': index_only, 'OVERWRITE': overwrite, 'DEPTH': depth, 'PLUGINS': plugins or '', 'DEFAULT_PERSONA': persona or 'Default', } # Merge custom config overrides config.update(custom_config) crawl = Crawl.objects.create( urls=urls_content, max_depth=depth, tags_str=tag, notes=notes, label=f'{self.request.user.username}@{HOSTNAME}{self.request.path} {timestamp}', created_by_id=self.request.user.pk, config=config ) # 3. create a CrawlSchedule if schedule is provided if schedule: from archivebox.crawls.models import CrawlSchedule crawl_schedule = CrawlSchedule.objects.create( template=crawl, schedule=schedule, is_enabled=True, label=crawl.label, notes=f"Auto-created from add page. {notes}".strip(), created_by_id=self.request.user.pk, ) crawl.schedule = crawl_schedule crawl.save(update_fields=['schedule']) # 4. start the Orchestrator & wait until it completes # ... orchestrator will create the root Snapshot, which creates pending ArchiveResults, which gets run by the ArchiveResultActors ... # from archivebox.crawls.actors import CrawlActor # from archivebox.core.actors import SnapshotActor, ArchiveResultActor rough_url_count = urls.count('://') # Build success message with schedule link if created schedule_msg = "" if schedule: schedule_msg = f" and scheduled to repeat {schedule}" messages.success( self.request, mark_safe(f"Created crawl with {rough_url_count} starting URL(s){schedule_msg}. Snapshots will be created and archived in the background. View Crawl →"), ) # Orchestrator (managed by supervisord) will pick up the queued crawl return redirect(crawl.admin_change_url) class HealthCheckView(View): """ A Django view that renders plain text "OK" for service discovery tools """ def get(self, request): """ Handle a GET request """ return HttpResponse( 'OK', content_type='text/plain', status=200 ) import json from django.http import JsonResponse def live_progress_view(request): """Simple JSON endpoint for live progress status - used by admin progress monitor.""" try: from archivebox.workers.orchestrator import Orchestrator from archivebox.crawls.models import Crawl from archivebox.core.models import Snapshot, ArchiveResult from django.db.models import Case, When, Value, IntegerField # Get orchestrator status orchestrator_running = Orchestrator.is_running() total_workers = Orchestrator().get_total_worker_count() if orchestrator_running else 0 # Get model counts by status crawls_pending = Crawl.objects.filter(status=Crawl.StatusChoices.QUEUED).count() crawls_started = Crawl.objects.filter(status=Crawl.StatusChoices.STARTED).count() # Get recent crawls (last 24 hours) from datetime import timedelta one_day_ago = timezone.now() - timedelta(days=1) crawls_recent = Crawl.objects.filter(created_at__gte=one_day_ago).count() snapshots_pending = Snapshot.objects.filter(status=Snapshot.StatusChoices.QUEUED).count() snapshots_started = Snapshot.objects.filter(status=Snapshot.StatusChoices.STARTED).count() archiveresults_pending = ArchiveResult.objects.filter(status=ArchiveResult.StatusChoices.QUEUED).count() archiveresults_started = ArchiveResult.objects.filter(status=ArchiveResult.StatusChoices.STARTED).count() archiveresults_succeeded = ArchiveResult.objects.filter(status=ArchiveResult.StatusChoices.SUCCEEDED).count() archiveresults_failed = ArchiveResult.objects.filter(status=ArchiveResult.StatusChoices.FAILED).count() # Get recently completed ArchiveResults with thumbnails (last 20 succeeded results) recent_thumbnails = [] recent_results = ArchiveResult.objects.filter( status=ArchiveResult.StatusChoices.SUCCEEDED, ).select_related('snapshot').order_by('-end_ts')[:20] for ar in recent_results: embed = ar.embed_path() if embed: # Only include results with embeddable image/media files ext = embed.lower().split('.')[-1] if '.' in embed else '' is_embeddable = ext in ('png', 'jpg', 'jpeg', 'gif', 'webp', 'svg', 'ico', 'pdf', 'html') if is_embeddable or ar.plugin in ('screenshot', 'favicon', 'dom'): recent_thumbnails.append({ 'id': str(ar.id), 'plugin': ar.plugin, 'snapshot_id': str(ar.snapshot_id), 'snapshot_url': ar.snapshot.url[:60] if ar.snapshot else '', 'embed_path': embed, 'archive_path': f'/archive/{ar.snapshot.timestamp}/{embed}' if ar.snapshot else '', 'end_ts': ar.end_ts.isoformat() if ar.end_ts else None, }) # Build hierarchical active crawls with nested snapshots and archive results from django.db.models import Prefetch active_crawls_qs = Crawl.objects.filter( status__in=[Crawl.StatusChoices.QUEUED, Crawl.StatusChoices.STARTED] ).prefetch_related( 'snapshot_set', 'snapshot_set__archiveresult_set', ).distinct().order_by('-modified_at')[:10] active_crawls = [] for crawl in active_crawls_qs: # Get ALL snapshots for this crawl to count status (already prefetched) all_crawl_snapshots = list(crawl.snapshot_set.all()) # Count snapshots by status from ALL snapshots total_snapshots = len(all_crawl_snapshots) completed_snapshots = sum(1 for s in all_crawl_snapshots if s.status == Snapshot.StatusChoices.SEALED) started_snapshots = sum(1 for s in all_crawl_snapshots if s.status == Snapshot.StatusChoices.STARTED) pending_snapshots = sum(1 for s in all_crawl_snapshots if s.status == Snapshot.StatusChoices.QUEUED) # Get only ACTIVE snapshots to display (limit to 5 most recent) active_crawl_snapshots = [ s for s in all_crawl_snapshots if s.status in [Snapshot.StatusChoices.QUEUED, Snapshot.StatusChoices.STARTED] ][:5] # Count URLs in the crawl (for when snapshots haven't been created yet) urls_count = 0 if crawl.urls: urls_count = len([u for u in crawl.urls.split('\n') if u.strip() and not u.startswith('#')]) # Calculate crawl progress crawl_progress = int((completed_snapshots / total_snapshots) * 100) if total_snapshots > 0 else 0 # Get active snapshots for this crawl (already prefetched) active_snapshots_for_crawl = [] for snapshot in active_crawl_snapshots: # Get archive results for this snapshot (already prefetched) snapshot_results = snapshot.archiveresult_set.all() # Count in memory instead of DB queries total_plugins = len(snapshot_results) completed_plugins = sum(1 for ar in snapshot_results if ar.status == ArchiveResult.StatusChoices.SUCCEEDED) failed_plugins = sum(1 for ar in snapshot_results if ar.status == ArchiveResult.StatusChoices.FAILED) pending_plugins = sum(1 for ar in snapshot_results if ar.status == ArchiveResult.StatusChoices.QUEUED) # Calculate snapshot progress snapshot_progress = int(((completed_plugins + failed_plugins) / total_plugins) * 100) if total_plugins > 0 else 0 # Get all extractor plugins for this snapshot (already prefetched, sort in Python) # Order: started first, then queued, then completed def plugin_sort_key(ar): status_order = { ArchiveResult.StatusChoices.STARTED: 0, ArchiveResult.StatusChoices.QUEUED: 1, ArchiveResult.StatusChoices.SUCCEEDED: 2, ArchiveResult.StatusChoices.FAILED: 3, } return (status_order.get(ar.status, 4), ar.plugin) all_plugins = [ { 'id': str(ar.id), 'plugin': ar.plugin, 'status': ar.status, } for ar in sorted(snapshot_results, key=plugin_sort_key) ] active_snapshots_for_crawl.append({ 'id': str(snapshot.id), 'url': snapshot.url[:80], 'status': snapshot.status, 'started': snapshot.modified_at.isoformat() if snapshot.modified_at else None, 'progress': snapshot_progress, 'total_plugins': total_plugins, 'completed_plugins': completed_plugins, 'failed_plugins': failed_plugins, 'pending_plugins': pending_plugins, 'all_plugins': all_plugins, }) # Check if crawl can start (for debugging stuck crawls) can_start = bool(crawl.urls) urls_preview = crawl.urls[:60] if crawl.urls else None # Check if retry_at is in the future (would prevent worker from claiming) retry_at_future = crawl.retry_at > timezone.now() if crawl.retry_at else False seconds_until_retry = int((crawl.retry_at - timezone.now()).total_seconds()) if crawl.retry_at and retry_at_future else 0 active_crawls.append({ 'id': str(crawl.id), 'label': str(crawl)[:60], 'status': crawl.status, 'started': crawl.modified_at.isoformat() if crawl.modified_at else None, 'progress': crawl_progress, 'max_depth': crawl.max_depth, 'urls_count': urls_count, 'total_snapshots': total_snapshots, 'completed_snapshots': completed_snapshots, 'started_snapshots': started_snapshots, 'failed_snapshots': 0, 'pending_snapshots': pending_snapshots, 'active_snapshots': active_snapshots_for_crawl, 'can_start': can_start, 'urls_preview': urls_preview, 'retry_at_future': retry_at_future, 'seconds_until_retry': seconds_until_retry, }) return JsonResponse({ 'orchestrator_running': orchestrator_running, 'total_workers': total_workers, 'crawls_pending': crawls_pending, 'crawls_started': crawls_started, 'crawls_recent': crawls_recent, 'snapshots_pending': snapshots_pending, 'snapshots_started': snapshots_started, 'archiveresults_pending': archiveresults_pending, 'archiveresults_started': archiveresults_started, 'archiveresults_succeeded': archiveresults_succeeded, 'archiveresults_failed': archiveresults_failed, 'active_crawls': active_crawls, 'recent_thumbnails': recent_thumbnails, 'server_time': timezone.now().isoformat(), }) except Exception as e: import traceback return JsonResponse({ 'error': str(e), 'traceback': traceback.format_exc(), 'orchestrator_running': False, 'total_workers': 0, 'crawls_pending': 0, 'crawls_started': 0, 'crawls_recent': 0, 'snapshots_pending': 0, 'snapshots_started': 0, 'archiveresults_pending': 0, 'archiveresults_started': 0, 'archiveresults_succeeded': 0, 'archiveresults_failed': 0, 'active_crawls': [], 'recent_thumbnails': [], 'server_time': timezone.now().isoformat(), }, status=500) def find_config_section(key: str) -> str: CONFIGS = get_all_configs() if key in CONSTANTS_CONFIG: return 'CONSTANT' matching_sections = [ section_id for section_id, section in CONFIGS.items() if key in dict(section) ] section = matching_sections[0] if matching_sections else 'DYNAMIC' return section def find_config_default(key: str) -> str: CONFIGS = get_all_configs() if key in CONSTANTS_CONFIG: return str(CONSTANTS_CONFIG[key]) default_val = None for config in CONFIGS.values(): if key in dict(config): default_field = getattr(config, 'model_fields', dict(config))[key] default_val = default_field.default if hasattr(default_field, 'default') else default_field break if isinstance(default_val, Callable): default_val = inspect.getsource(default_val).split('lambda', 1)[-1].split(':', 1)[-1].replace('\n', ' ').strip() if default_val.count(')') > default_val.count('('): default_val = default_val[:-1] else: default_val = str(default_val) return default_val def find_config_type(key: str) -> str: from typing import get_type_hints, ClassVar CONFIGS = get_all_configs() for config in CONFIGS.values(): if hasattr(config, key): # Try to get from pydantic model_fields first (more reliable) if hasattr(config, 'model_fields') and key in config.model_fields: field = config.model_fields[key] if hasattr(field, 'annotation'): try: return str(field.annotation.__name__) except AttributeError: return str(field.annotation) # Fallback to get_type_hints with proper namespace try: import typing namespace = { 'ClassVar': ClassVar, 'Optional': typing.Optional, 'Union': typing.Union, 'List': typing.List, 'Dict': typing.Dict, 'Path': Path, } type_hints = get_type_hints(config, globalns=namespace, localns=namespace) try: return str(type_hints[key].__name__) except AttributeError: return str(type_hints[key]) except Exception: # If all else fails, return str pass return 'str' def key_is_safe(key: str) -> bool: for term in ('key', 'password', 'secret', 'token'): if term in key.lower(): return False return True def find_config_source(key: str, merged_config: dict) -> str: """Determine where a config value comes from.""" import os from archivebox.machine.models import Machine # Check if it's from archivebox.machine.config try: machine = Machine.current() if machine.config and key in machine.config: return 'Machine' except Exception: pass # Check if it's from environment variable if key in os.environ: return 'Environment' # Check if it's from archivebox.config.file from archivebox.config.configset import BaseConfigSet file_config = BaseConfigSet.load_from_file(CONSTANTS.CONFIG_FILE) if key in file_config: return 'Config File' # Otherwise it's using the default return 'Default' @render_with_table_view def live_config_list_view(request: HttpRequest, **kwargs) -> TableContext: CONFIGS = get_all_configs() assert request.user.is_superuser, 'Must be a superuser to view configuration settings.' # Get merged config that includes Machine.config overrides try: from archivebox.machine.models import Machine machine = Machine.current() merged_config = get_config() except Exception as e: # Fallback if Machine model not available merged_config = get_config() machine = None rows = { "Section": [], "Key": [], "Type": [], "Value": [], "Source": [], "Default": [], # "Documentation": [], # "Aliases": [], } for section_id, section in reversed(list(CONFIGS.items())): for key in dict(section).keys(): rows['Section'].append(section_id) # section.replace('_', ' ').title().replace(' Config', '') rows['Key'].append(ItemLink(key, key=key)) rows['Type'].append(format_html('{}', find_config_type(key))) # Use merged config value (includes machine overrides) actual_value = merged_config.get(key, getattr(section, key, None)) rows['Value'].append(mark_safe(f'{actual_value}') if key_is_safe(key) else '******** (redacted)') # Show where the value comes from source = find_config_source(key, merged_config) source_colors = { 'Machine': 'purple', 'Environment': 'blue', 'Config File': 'green', 'Default': 'gray' } rows['Source'].append(format_html('{}', source_colors.get(source, 'gray'), source)) rows['Default'].append(mark_safe(f'{find_config_default(key) or "See here..."}')) # rows['Documentation'].append(mark_safe(f'Wiki: {key}')) # rows['Aliases'].append(', '.join(find_config_aliases(key))) section = 'CONSTANT' for key in CONSTANTS_CONFIG.keys(): rows['Section'].append(section) # section.replace('_', ' ').title().replace(' Config', '') rows['Key'].append(ItemLink(key, key=key)) rows['Type'].append(format_html('{}', getattr(type(CONSTANTS_CONFIG[key]), '__name__', str(CONSTANTS_CONFIG[key])))) rows['Value'].append(format_html('{}', CONSTANTS_CONFIG[key]) if key_is_safe(key) else '******** (redacted)') rows['Source'].append(mark_safe('Constant')) rows['Default'].append(mark_safe(f'{find_config_default(key) or "See here..."}')) # rows['Documentation'].append(mark_safe(f'Wiki: {key}')) # rows['Aliases'].append('') return TableContext( title="Computed Configuration Values", table=rows, ) @render_with_item_view def live_config_value_view(request: HttpRequest, key: str, **kwargs) -> ItemContext: import os from archivebox.machine.models import Machine from archivebox.config.configset import BaseConfigSet CONFIGS = get_all_configs() FLAT_CONFIG = get_flat_config() assert request.user.is_superuser, 'Must be a superuser to view configuration settings.' # Get merged config merged_config = get_config() # Determine all sources for this config value sources_info = [] # Default value default_val = find_config_default(key) if default_val: sources_info.append(('Default', default_val, 'gray')) # Config file value if CONSTANTS.CONFIG_FILE.exists(): file_config = BaseConfigSet.load_from_file(CONSTANTS.CONFIG_FILE) if key in file_config: sources_info.append(('Config File', file_config[key], 'green')) # Environment variable if key in os.environ: sources_info.append(('Environment', os.environ[key] if key_is_safe(key) else '********', 'blue')) # Machine config machine = None machine_admin_url = None try: machine = Machine.current() machine_admin_url = f'/admin/machine/machine/{machine.id}/change/' if machine.config and key in machine.config: sources_info.append(('Machine', machine.config[key] if key_is_safe(key) else '********', 'purple')) except Exception: pass # Final computed value final_value = merged_config.get(key, FLAT_CONFIG.get(key, CONFIGS.get(key, None))) if not key_is_safe(key): final_value = '********' # Build sources display sources_html = '
'.join([ f'{source}: {value}' for source, value, color in sources_info ]) # aliases = USER_CONFIG.get(key, {}).get("aliases", []) aliases = [] if key in CONSTANTS_CONFIG: section_header = mark_safe(f'[CONSTANTS]   {key}   (read-only, hardcoded by ArchiveBox)') elif key in FLAT_CONFIG: section_header = mark_safe(f'data / ArchiveBox.conf   [{find_config_section(key)}]   {key}') else: section_header = mark_safe(f'[DYNAMIC CONFIG]   {key}   (read-only, calculated at runtime)') return ItemContext( slug=key, title=key, data=[ { "name": section_header, "description": None, "fields": { 'Key': key, 'Type': find_config_type(key), 'Value': final_value, 'Source': find_config_source(key, merged_config), }, "help_texts": { 'Key': mark_safe(f''' Documentation   Aliases: {", ".join(aliases)} '''), 'Type': mark_safe(f''' See full definition in archivebox/config... '''), 'Value': mark_safe(f''' {'Value is redacted for your security. (Passwords, secrets, API tokens, etc. cannot be viewed in the Web UI)

' if not key_is_safe(key) else ''}


Configuration Sources (in priority order):

{sources_html}

To change this value, edit data/ArchiveBox.conf or run:

archivebox config --set {key}="{ val.strip("'") if (val := find_config_default(key)) else (str(FLAT_CONFIG[key] if key_is_safe(key) else '********')).strip("'") }"

'''), 'Source': mark_safe(f''' The value shown in the "Value" field comes from the {find_config_source(key, merged_config)} source.

Priority order (highest to lowest):
  1. Machine - Machine-specific overrides (e.g., resolved binary paths) {f'
    → Edit {key} in Machine.config for this server' if machine_admin_url else ''}
  2. Environment - Environment variables
  3. Config File - data/ArchiveBox.conf
  4. Default - Default value from code
{f'
💡 Tip: To override {key} on this machine, edit the Machine.config field and add:
{{"\\"{key}\\": "your_value_here"}}' if machine_admin_url and key not in CONSTANTS_CONFIG else ''} '''), }, }, ], )