Skip to content
/ urlsup Public

✅ CLI to find broken URLs in files (awesome_bot alternative but significantly faster)

License

Notifications You must be signed in to change notification settings

simeg/urlsup

Repository files navigation

urlsup CI Code Coverage

urlsup (urls up) finds URLs in files and checks whether they are up by making a GET request and checking the response status code. This tool is useful for lists, repos or any type of project containing URLs that you want to be up.

It's written in Rust (stable) and executes the requests async in multiple threads, making it very fast. Uses browser-like HTTP client behavior with automatic protocol negotiation and reliable connection handling. This in combination with its ease of use makes it the perfect tool for your CI pipeline.

⚡️ A modern and significantly faster alternative to awesome_bot ⚡️

Use the GitHub Action urlsup-action to quickly get your CI pipeline up and running.

Using awesome_bot today? Here's a migration guide.

Dotfiles Banner

🎉 What's New in v2.0

urlsup v2.0 introduces a modern CLI design with breaking changes for better usability:

🔄 Renamed Flags (Breaking Changes)

  • --white-list--allowlist (modern terminology)
  • --allow--allow-status (clearer naming)
  • --threads--concurrency (industry standard)
  • --file-types--include (shorter, clearer)

✨ New Features

  • 📄 Configuration Files: TOML-based config with automatic discovery
  • 📤 Output Formats: JSON support for automation (--format json)
  • 📊 Progress Reporting: Beautiful progress bars with real-time stats
  • 🔍 Advanced Filtering: Regex-based URL exclusion patterns
  • 🔄 Retry Logic: Configurable retry attempts with exponential backoff
  • ⏱️ Rate Limiting: Built-in request throttling
  • 🔇 Quiet/Verbose Modes: Better control over output verbosity
  • 🚨 Enhanced Error Handling: Comprehensive error types with context
  • ⚡ Browser-like Behavior: HTTP client that behaves like web browsers for maximum compatibility
  • 📈 Performance Analysis: Built-in memory monitoring and optimization suggestions
  • 📊 HTML Dashboard: Rich visual reporting with charts and performance metrics

📚 Table of Contents

🚀 Usage

urlsup - CLI to validate URLs in files [version 2.4.0]

Usage: urlsup [OPTIONS] [FILES]... [COMMAND]

Example:

	$ urlsup . --recursive --include md,txt

Commands:
  completion-generate                Generate shell completions
  completion-install                 Install shell completions to standard location
  config-wizard                      Run interactive configuration wizard
  help                               Print this message or the help of the given subcommand(s)

Core Options:
  -r, --recursive                    Recursively process directories. Will skip files/directories listed in .gitignore
  -t, --timeout <SECONDS>            Connection timeout in seconds (default: 5)
      --concurrency <COUNT>          Concurrent requests (default: CPU cores)

Filtering & Content:
      --include <EXTENSIONS>         File extensions to process (e.g., md,html,txt)
      --allowlist <URLS>             URLs to allow (comma-separated)
      --allow-status <CODES>         Status codes to allow (comma-separated)
      --exclude-pattern <REGEX>      URL patterns to exclude (regex)

Retry & Rate Limiting:
      --retry <COUNT>                Retry attempts for failed requests (default: 0)
      --retry-delay <MS>             Delay between retries in ms (default: 1000)
      --rate-limit <MS>              Delay between requests in ms (default: 0)
      --allow-timeout                Allow URLs that timeout
      --failure-threshold <PERCENT>  Fail only if more than X% URLs are broken (0-100)

Output & Verbosity:
  -q, --quiet                        Suppress progress output
  -v, --verbose                      Enable verbose logging
      --format <FORMAT>              Output format [default: text] [text|json|minimal]
      --no-progress                  Disable progress bars

Network & Security:
      --user-agent <AGENT>           Custom User-Agent header
      --proxy <URL>                  HTTP/HTTPS proxy URL
      --insecure                     Skip SSL certificate verification

Configuration:
      --config <FILE>                Use specific config file
      --no-config                    Ignore config files

Performance Analysis:
      --show-performance             Show memory usage and optimization suggestions
      --html-dashboard <PATH>        Generate HTML dashboard report

Options:
  -h, --help                         Print help
  -V, --version                      Print version

📝 Examples

Basic File Checking

# Check a single file
$ urlsup README.md

# Check multiple files
$ urlsup README.md CHANGELOG.md

# Check files with wildcards
$ urlsup docs/*.md

Directory Processing

Important: urlsup treats files and directories differently:

  • Files: Directly processed (e.g., urlsup README.md)
  • Directories: Must use --recursive flag (e.g., urlsup --recursive docs/). Will skip files/directories listed in .gitignore.
# ❌ This will fail with an error
$ urlsup docs/
error: 'docs/' is a directory. Use --recursive to process directories.

# ✅ Process all files in a directory recursively
$ urlsup --recursive docs/

# ✅ Process only specific file types
$ urlsup --recursive --include md,txt docs/

# ✅ Process current directory recursively (will skip files/directories listed in .gitignore)
$ urlsup --recursive .

File Type Filtering

# Only check markdown and text files
$ urlsup --recursive --include md,txt .

# Only check web files
$ urlsup --recursive --include html,css,js website/

# Multiple extensions
$ urlsup --recursive --include md,rst,txt docs/

Advanced Options

# Allow specific status codes
$ urlsup README.md --allow-status 403,429

# Set timeout and allow timeouts
$ urlsup README.md --allow-timeout -t 5

# Allowlist URLs (partial matches)
$ urlsup README.md --allowlist rust,crates

# Failure threshold - only fail if more than X% of URLs are broken
$ urlsup --recursive docs/ --failure-threshold 10  # Allow up to 10% failures
$ urlsup README.md --failure-threshold 0           # Strict mode - fail on any broken URL

# Combine recursive with filtering and options
$ urlsup --recursive --include md --allow-status 403 --timeout 10 docs/

# Use quiet mode for scripts
$ urlsup --quiet --recursive docs/

# Enable verbose output for debugging
$ urlsup --verbose README.md

# Use JSON output format
$ urlsup --format json README.md

# Exclude URLs with patterns
$ urlsup --exclude-pattern ".*\.local$" --exclude-pattern "^http://localhost.*" docs/

# Performance analysis and reporting
$ urlsup --show-performance README.md
$ urlsup --html-dashboard report.html docs/

Git Integration

When using --recursive, urlsup automatically respects your .gitignore files:

# This will skip files/directories listed in .gitignore
$ urlsup --recursive .

# Examples of automatically ignored paths:
# - node_modules/
# - target/
# - .git/
# - *.log files
# - Any patterns in your .gitignore

This means you don't need to manually exclude build artifacts, dependencies, or other generated files.

📦 Installation

Install with cargo to run urlsup on your local machine.

cargo install urlsup

🚀 Shell Completions Installation

urlsup supports shell completions for bash, zsh, and fish. You can generate completions manually or use the built-in installation command for automatic setup.

Automatic Installation (Recommended)

The completion-install command automatically installs shell completions to standard directories and provides setup instructions:

# Install bash completions
$ urlsup completion-install bash
✅ Shell completions installed successfully!

Completion installed to: /Users/user/.local/share/bash-completion/completions/urlsup

To enable bash completions, add this to your ~/.bashrc or ~/.bash_profile:
if [[ -d ~/.local/share/bash-completion/completions ]]; then
    for completion in ~/.local/share/bash-completion/completions/*; do
        [[ -r "$completion" ]] && source "$completion"
    done
fi

Then restart your shell or run: source ~/.bashrc

# Install zsh completions
$ urlsup completion-install zsh
✅ Shell completions installed successfully!

Completion installed to: /Users/user/.local/share/zsh/site-functions/_urlsup

To enable zsh completions, add this to your ~/.zshrc:
if [[ -d ~/.local/share/zsh/site-functions ]]; then
    fpath=(~/.local/share/zsh/site-functions $fpath)
    autoload -U compinit && compinit
fi

Then restart your shell or run: source ~/.zshrc
You may also need to clear the completion cache: rm -f ~/.zcompdump*

# Install fish completions
$ urlsup completion-install fish
✅ Shell completions installed successfully!

Completion installed to: /Users/user/.config/fish/completions/urlsup.fish

Fish completions are automatically loaded from ~/.config/fish/completions/
Restart your shell or run: fish -c 'complete --erase; source ~/.config/fish/config.fish'

Manual Installation

For manual installation or unsupported shells, generate the completion script and add it yourself:

# Generate completions for your shell
$ urlsup completion-generate bash > urlsup_completion.bash
$ urlsup completion-generate zsh > _urlsup
$ urlsup completion-generate fish > urlsup.fish
$ urlsup completion-generate powershell > urlsup_completion.ps1
$ urlsup completion-generate elvish > urlsup_completion.elv

# Then add to your shell's configuration manually

Supported Shells

Shell Auto-Install Manual Install Standard Location
bash ✅ Yes ✅ Yes ~/.local/share/bash-completion/completions/urlsup
zsh ✅ Yes ✅ Yes ~/.local/share/zsh/site-functions/_urlsup
fish ✅ Yes ✅ Yes ~/.config/fish/completions/urlsup.fish
PowerShell ❌ Manual only ✅ Yes Add to $PROFILE manually
Elvish ❌ Manual only ✅ Yes Add to ~/.elvish/rc.elv manually

Note: The completion-install command creates directories as needed and handles path resolution automatically. For PowerShell and Elvish, use the manual completion-generate command and follow the provided instructions.

🧙‍♂️ Configuration Wizard

For new users or complex setups, urlsup includes an interactive configuration wizard that guides you through creating optimal configurations:

# Run the interactive setup wizard
$ urlsup config-wizard

🧙‍♂️ urlsup Configuration Wizard
Let's set up urlsup for your project!

📋 What type of project are you setting up?
  Documentation Site
  GitHub Repository
  Blog/Content Site
  API Documentation
  Wiki/Knowledge Base
  CI/CD Pipeline
> Custom Setup

Available Project Templates:

Template Description Optimized For
Documentation Site Static site generators (Jekyll, Hugo, Gatsby) Fast builds, external links
GitHub Repository README, contributing guides, documentation Community projects, mixed URLs
Blog/Content Site WordPress, articles, content management External content, social links
API Documentation OpenAPI specs, endpoint documentation API reliability, auth handling
Wiki/Knowledge Base Internal wikis, documentation hubs Internal links, team wikis
CI/CD Pipeline Automated validation in CI Fast execution, strict validation
Custom Setup Configure everything manually Full control over all settings

Wizard Features:

  • 📋 Smart Templates: Pre-configured settings for common use cases
  • 🎯 Interactive Filtering: Set up allowlists and exclusion patterns
  • ⚙️ Advanced Configuration: Timeout, retry, and performance tuning
  • 💾 Automatic Generation: Creates .urlsup.toml with explanatory comments
  • 📚 Usage Examples: Shows relevant commands for your configuration

The wizard creates optimized configurations with appropriate defaults for timeouts, retry logic, file types, and failure thresholds based on your project type.

⚙️ Configuration File

urlsup supports TOML configuration files for managing complex setups. Place a .urlsup.toml file in your project root (or use the wizard above to generate one automatically):

# .urlsup.toml - Project configuration for urlsup
timeout = 5
threads = 8  # Number of concurrent threads (maps to --concurrency CLI option)
allow_timeout = false
file_types = ["md", "html", "txt"]

# URL patterns to exclude (regex)
exclude_patterns = [
    "^https://example\\.com/private/.*",
    ".*\\.local$",
    "^http://localhost.*"
]

# URLs to allowlist
allowlist = [
    "https://api.github.com",
    "https://docs.rs"
]

# HTTP status codes to allow
allowed_status_codes = [403, 429]

# Advanced network settings
user_agent = "MyBot/1.0"
retry_attempts = 3
retry_delay = 1000  # milliseconds
rate_limit_delay = 100  # milliseconds between requests
failure_threshold = 10.0  # Allow up to 10% of URLs to fail

# Performance settings
use_head_requests = false  # Use HEAD instead of GET for faster validation

# Security settings
skip_ssl_verification = false
proxy = "http://proxy.company.com:8080"

# Output settings
output_format = "text"  # or "json" or "minimal"
verbose = false

# Performance analysis
show_performance = false  # Show memory usage and optimization suggestions

Configuration Discovery

urlsup searches for configuration files in this order:

  1. .urlsup.toml in current directory
  2. .urlsup.toml in parent directories (up to 3 levels)
  3. Default configuration if no file found

CLI arguments always override configuration file settings.

🔧 Advanced Features

🎯 Failure Threshold

Control when urlsup should fail based on the percentage of broken URLs:

# Only fail if more than 20% of URLs are broken
$ urlsup --recursive docs/ --failure-threshold 20

# Strict mode - fail on any broken URL (default behavior)
$ urlsup docs/ --failure-threshold 0

# Lenient mode for large documentation sets
$ urlsup --recursive . --failure-threshold 5  # Allow up to 5% failures

Configuration file:

# In .urlsup.toml
failure_threshold = 10.0  # Allow up to 10% failures

Use Cases:

  • Large documentation: Prevent CI failures for 1-2 stale external links out of hundreds
  • External API monitoring: Allow some endpoints to be temporarily down
  • Migration periods: Gradually improve link quality without breaking builds
  • Third-party content: Handle external links that may be occasionally unreachable

Example output:

$ urlsup --recursive docs/ --failure-threshold 15

# When within threshold:
✅ Failure rate 12.5% is within threshold 15.0% (5/40 URLs failed)

# When exceeding threshold:
❌ Failure rate 17.5% exceeds threshold 15.0% (7/40 URLs failed)

Retry Logic & Rate Limiting

Handle flaky networks and respect server limits:

# Configure via CLI (basic)
$ urlsup --timeout 60 README.md

# Configure via .urlsup.toml (advanced)
retry_attempts = 5
retry_delay = 2000
rate_limit_delay = 500

URL Exclusion Patterns

Exclude URLs matching regex patterns:

# In .urlsup.toml
exclude_patterns = [
    "^https://internal\\.company\\.com/.*",  # Skip internal URLs
    ".*\\.local$",                           # Skip .local domains
    "^http://localhost.*",                   # Skip localhost
    "https://example\\.com/api/.*"           # Skip API endpoints
]

📊 Progress Reporting

Beautiful progress bars for large operations:

# Progress bars are enabled automatically for TTY terminals
$ urlsup --recursive docs/

# Output includes:
# ⠋ [00:01:23] [████████████████████████████████████████] 150/150 files processed
# ⠙ [00:00:45] [██████████████████████████              ] 245/320 URLs validated (76% successful)

🌐 Custom User Agent & Proxy Support

# In .urlsup.toml
user_agent = "MyCompany/URLChecker 2.0"
proxy = "http://proxy.company.com:8080"
skip_ssl_verification = false  # Set to true for internal/dev environments

📤 Output Formats

# Text output (default) - clean, colorful, emoji-based with grouping
$ urlsup README.md
✅ No issues found!

# JSON output for scripts and automation
$ urlsup --format json README.md
{"status": "success", "issues": []}

# Minimal output for scripts (no colors, emojis, or config info)
$ urlsup --format minimal README.md
404 https://example.com/broken
500 https://api.broken.com

# Or configure in .urlsup.toml
output_format = "json"

📊 JSON Output Examples

JSON format is perfect for automation, CI/CD integration, and programmatic processing:

Successful validation:

$ urlsup --format json README.md
{"status": "success", "issues": []}

Failed validation with issues:

$ urlsup --format json docs/
{"status": "failure", "issues": [
  {"url": "https://example.com/404", "file": "docs/api.md", "line": 23, "status_code": 404, "description": ""},
  {"url": "https://broken.link", "file": "docs/guide.md", "line": 45, "status_code": null, "description": "connection timeout"}
]}

Processing JSON with jq:

# Extract all broken URLs
$ urlsup --format json docs/ | jq -r '.issues[].url'

# Count issues by status code
$ urlsup --format json docs/ | jq '.issues | group_by(.status_code) | map({status: .[0].status_code, count: length})'

# Find all timeout errors
$ urlsup --format json docs/ | jq '.issues[] | select(.description | contains("timeout"))'

# Get files with broken links
$ urlsup --format json docs/ | jq -r '.issues[].file' | sort | uniq

# Export issues to CSV for reporting
$ urlsup --format json docs/ | jq -r '.issues[] | [.file, .line, .url, .status_code] | @csv'

Output Format Comparison

Format Colors/Emojis Config Info URL List Progress Bars Issue Grouping Use Case
text ✅ Yes ✅ Yes ✅ Yes ✅ Yes ✅ Yes Interactive use
json ❌ No ❌ No ❌ No ❌ No ❌ No Automation/scripts
minimal ❌ No ❌ No ❌ No ❌ No ❌ No Simple scripts/CI

Verbose Logging

# Enable verbose output via CLI
$ urlsup --verbose README.md

# Or configure in .urlsup.toml
verbose = true

# Quiet mode for scripts (minimal output)
$ urlsup --quiet README.md

Verbose mode provides detailed information about:

  • Files being processed
  • URLs found and filtered
  • Request progress and timing
  • Configuration settings used

📈 Performance Analysis

Get detailed insights into memory usage and performance characteristics:

# Enable performance monitoring with optimization suggestions
$ urlsup --show-performance README.md

# Example output:
# ⚡ Performance Analysis
# Total execution time: 2.34s
# Peak memory usage: 45.2 MB
# Average CPU usage: 23.4%
# 
# 📊 Operation Breakdown:
# • File processing: 0.12s (156 files)
# • URL discovery: 0.89s (1,247 URLs found)
# • URL validation: 1.33s (987 unique URLs validated)
# 
# 💡 Optimization Suggestions:
# • Consider using --concurrency 8 for better performance
# • Enable HEAD requests for faster validation (use_head_requests = true)
# • Add .gitignore patterns to reduce file processing overhead

Configuration:

# In .urlsup.toml
show_performance = true  # Always show performance analysis

Use Cases:

  • Performance tuning: Identify bottlenecks in large documentation sets
  • CI/CD optimization: Monitor resource usage in automated pipelines
  • Capacity planning: Understand resource requirements for scaling
  • Troubleshooting: Debug slow validation issues

🎨 Adaptive Color Schemes

urlsup automatically detects your terminal's theme and adapts its color scheme for optimal readability:

Smart Terminal Detection:

  • Light Themes: Uses dark text colors with high contrast for readability
  • Dark Themes: Uses bright colors optimized for dark backgrounds
  • Unknown Themes: Falls back to universally compatible colors

Detection Methods:

  • Environment variables (URLSUP_THEME, COLORFGBG, TERM_PROGRAM)
  • Terminal program detection (iTerm, VS Code, Apple Terminal, etc.)
  • SSH context detection (defaults to dark theme for remote sessions)
  • True color support detection for enhanced color rendering

Manual Override:

# Force a specific theme
export URLSUP_THEME=light    # or 'dark' or 'unknown'
urlsup README.md

# The tool will adapt all colors accordingly

WCAG Accessibility Compliance:

  • All color combinations meet WCAG AA accessibility standards (4.5:1 contrast ratio)
  • Automatic fallback to high-contrast colors when needed
  • Semantic color meanings remain consistent across themes

This ensures that urlsup looks great and remains readable regardless of your terminal setup, whether you're using a light IDE theme, dark terminal, or working over SSH.

📊 HTML Dashboard

Generate comprehensive visual reports with charts and detailed analysis:

# Generate HTML dashboard with performance metrics
$ urlsup --html-dashboard report.html --show-performance docs/

# Dashboard includes:
# • Interactive charts showing validation results
# • Performance metrics and timing breakdowns
# • Detailed issue listings with file locations
# • Configuration summary and recommendations
# • Responsive design for desktop and mobile viewing

Features:

  • 📊 Interactive Charts: Doughnut charts showing success/failure rates by category
  • 📈 Performance Metrics: Memory usage, CPU utilization, and timing analysis
  • 🔍 Detailed Issue Tracking: Line-by-line breakdown of broken URLs
  • 💡 Smart Recommendations: Optimization suggestions based on actual usage patterns
  • 📱 Responsive Design: Works perfectly on desktop, tablet, and mobile devices
  • 🎨 Modern UI: Clean, professional styling with dark/light theme support

Dashboard Sections:

  1. Executive Summary: Key metrics and success rates
  2. Validation Results: Interactive visualization of URL status distribution
  3. Performance Analysis: Detailed timing and resource usage breakdown
  4. Issue Details: Comprehensive list of broken URLs with context
  5. Optimization Recommendations: Actionable suggestions for improvement

Example Usage in CI/CD:

# .github/workflows/urls.yml
- name: Validate URLs and generate report
  run: |
    urlsup --html-dashboard validation-report.html --show-performance docs/
    
- name: Upload report artifact
  uses: actions/upload-artifact@v3
  with:
    name: url-validation-report
    path: validation-report.html

Sample Dashboard Output: The HTML dashboard provides a complete overview of your URL validation results with:

  • Visual charts showing the health of your URLs
  • Performance insights to optimize future runs
  • Detailed breakdown of any issues found
  • Professional presentation suitable for stakeholder reporting

🔒 Security Features

SSL Certificate Verification

# Skip SSL verification for internal/development URLs
skip_ssl_verification = true

⚠️ Warning: Only disable SSL verification for trusted internal environments.

Proxy Support

# HTTP/HTTPS proxy configuration
proxy = "http://username:[email protected]:8080"

Supports both HTTP and HTTPS proxies with optional authentication.

🚀 Performance Features

HEAD Request Optimization

For even faster URL validation, enable HEAD requests instead of GET requests:

# In .urlsup.toml
use_head_requests = true  # Use HEAD instead of GET for faster validation

Benefits:

  • Faster validation: HEAD requests only fetch headers, not full content
  • Reduced bandwidth: Minimal data transfer for each URL check
  • Better for CI/CD: Faster pipeline execution for large documentation sets

When to use:

  • ✅ Internal documentation validation
  • ✅ Known-good server environments
  • ✅ CI/CD pipelines with trusted URL sets
  • ✅ Large-scale validation where speed is critical

When NOT to use:

  • ❌ Public URL validation (some servers reject HEAD requests)
  • ❌ Mixed server environments with unknown HEAD support
  • ❌ First-time validation of unknown URLs

Example usage:

# Enable HEAD requests for faster CI validation
$ urlsup --config .urlsup-fast.toml --recursive docs/

# Where .urlsup-fast.toml contains:
use_head_requests = true
timeout = 15
threads = 16

⚡ Browser-like HTTP Client

urlsup uses a simplified HTTP client designed for maximum compatibility:

🌐 Browser-Compatible Behavior

  • Automatic Protocol Negotiation: Lets the client and server automatically negotiate HTTP/1.1 or HTTP/2
  • Default Connection Management: Uses reqwest's browser-like connection handling for reliability
  • Automatic Compression: Leverages gzip, brotli, and deflate for reduced bandwidth (like browsers)
  • Reliable Error Handling: Avoids complex optimizations that can cause connection issues

🎯 Memory & Algorithm Improvements

  • Ultra-Fast Hashing: Uses FxHashSet for 15-20% faster URL deduplication
  • Smart Pre-allocation: File-type-aware capacity estimation (Markdown 2x, HTML 3x multipliers)
  • Optimized Deduplication: O(n) hash-based instead of O(n²) sorting-based
  • Memory-Efficient Streaming: Handles large URL sets without memory bloat
  • Adaptive Sizing: Dynamic memory allocation based on file types and URL patterns
  • SIMD-Optimized String Processing: Uses memchr for vectorized URL pattern detection
  • Vectorized Line Processing: Chunked processing with cache-friendly memory access patterns

🔄 Concurrent Processing

  • Dynamic Batch Sizing: Batch sizes adapt to URL count and system resources (2-100 range)
  • Connection Pooling: Optimized HTTP connection reuse with configurable pool limits
  • Token Bucket Rate Limiting: Smooth request distribution vs simple delays
  • Batched Progress Updates: Reduced atomic operations for better concurrent performance
  • Static Resource Reuse: Eliminates repeated allocations for parsing components

📈 Performance Gains

  • Small workloads (10-100 URLs): 25-35% faster validation with optimized batch sizing
  • Large workloads (1000+ URLs): 45-65% faster with 60-80% less memory usage
  • Memory efficiency: File-type-aware allocation reduces memory waste by 30-50%
  • Network optimization: Connection pooling and token bucket rate limiting improve throughput
  • CI/CD pipelines: Dramatically reduced execution time for documentation validation

🚨 Error Handling

Comprehensive error handling with specific error types:

  • Configuration errors: Invalid TOML, missing files
  • Network errors: Timeouts, connection failures, DNS resolution
  • Path errors: Invalid file paths, permission issues
  • Validation errors: Malformed URLs, regex compilation failures

All errors include helpful context and suggestions for resolution.

🔄 GitHub Actions

See urlsup-action.

🛠️ Development

This repo uses a Makefile as an interface for common operations.

  1. Do code changes
  2. Run make build link to build the project and create a symlink from the built binary to the root of the project
  3. Run ./urlsup to execute the binary with your changes
  4. Profit ⭐

About

✅ CLI to find broken URLs in files (awesome_bot alternative but significantly faster)

Resources

License

Stars

Watchers

Forks

Languages