Initializing OpenClaw

Powered by OpenClaw AI

STOP TRUSTING GHOST CODE

Drop a GitHub link. Our AI tears it apart — finds the lies, exposes the larp, and delivers a verdict in seconds.

Paste any GitHub repository URL

openclaw_agent v0.1 -- beta
scroll
Fake Repo Detection Code Quality Analysis Commit History Forensics Dependency Audit Larp vs. Reality AI-Powered Verdict OpenClaw Engine Fake Repo Detection Code Quality Analysis Commit History Forensics Dependency Audit Larp vs. Reality AI-Powered Verdict OpenClaw Engine

01 -- Process

HOW THE
CLAW WORKS

The GitHub ecosystem is flooded with performative repos — all README, no code. ClawGit cuts through the noise with multi-layer AI analysis.

01

Drop the Link

Paste any GitHub repository URL. Public repos, private coming soon.

02

OpenClaw Digs In

Our agent scans commits, code structure, dependencies, tests, and contribution patterns.

03

Verdict Delivered

Legit or Fake Repo — with detailed breakdown, confidence score, and specific red flags.

github.com/cool-ai-startup/super-model
SCANNING
0commits scanned
0files analyzed
0deps checked
X
No test files found anywhere in repo. Zero coverage.
X
README references 98.7% accuracy — no benchmark code, no eval script.
!
87% of commits from single day. Suspicious burst pattern.
!
Model weights referenced but not included or linked.
+
License file present (MIT). Repo structure looks standard.
FAKE REPO
94%confidence

02 -- Capabilities

WHAT WE
TEAR APART

Deep Code Forensics

OpenClaw reads every file, not just README. We check if the claimed functionality actually exists in code — function definitions, imports, execution paths, and logical coherence.

Beta

Commit Pattern Analysis

Detect artificial commit inflation, single-day dumps, and copied histories.

Beta

Test Coverage Check

Legitimate software has tests. We scan for test files, coverage reports, CI configs. No tests = big red flag.

Live

Metric Verification

Claimed "99% accuracy"? We look for the eval code, datasets, and benchmarks that would produce that number.

Beta

Dependency Reality Check

Are claimed dependencies real, pinned, and actually used in code?

Coming Soon

AI Model Validation

For AI repos specifically — does the architecture code match the claimed paper? Are weights linked? Can it actually run? We cross-reference arXiv, HuggingFace, and run feasibility checks.

Coming Soon

Full Report Export Coming Soon

Shareable PDF reports with full verdict breakdown, evidence citations, and confidence scoring. Share with your team, investors, or post to call out the larp publicly.

DATA

03 -- Numbers

WHAT WE
FOUND SO FAR

0Repos Analyzedin closed beta
0%Had Zero Testsof "AI" repos scanned
0%Fake Verdict Ratebeta sample
0sAvg Scan Timetarget -- in optimization

Metrics from closed beta testing -- small sample size, subject to change

04 -- Interactive

ADJUST THE
TRUST THRESHOLD

Configure how strict ClawGit's verdict engine is. Higher threshold = more repos flagged.

Strictness Level50%
LenientSoftBalancedStrictParanoid
BALANCED MODE

05 -- Documentation

OPENCLAW AI
AGENT DOCS

What is ClawGit?

ClawGit is an AI-powered GitHub repository analysis platform designed to detect code fraud, performative programming, and misleading open-source projects. Powered by the OpenClaw engine, it performs deep multi-layer analysis of repository structure, code quality, commit patterns, and claimed metrics.

The Problem

The open-source ecosystem is increasingly polluted with repositories that look impressive on the surface but contain little to no real functionality. These "larp repos" feature polished READMEs, inflated star counts, and bold claims — but the actual code tells a different story. Developers waste time evaluating these projects, investors get misled, and the community suffers from noise drowning out genuine work.

The Solution

ClawGit uses AI-driven forensic analysis to cut through the noise. By examining every file, every commit, and every dependency, OpenClaw builds a comprehensive picture of a repository's legitimacy and delivers a confidence-weighted verdict backed by specific evidence.

Key Capabilities

  • Deep code forensics across every source file in the repository
  • Commit history pattern analysis and anomaly detection
  • Test infrastructure scanning and coverage estimation
  • Claimed metric verification against actual evaluation code
  • Dependency manifest validation and usage tracking
  • AI/ML model validation including architecture and weight checks
  • Confidence-weighted verdicts with detailed evidence reports

Current Status

ClawGit is currently in closed beta. The core scanning engine is operational with code forensics, commit analysis, and test coverage checking live. Dependency auditing and AI model validation are in active development. Full API access and PDF report exports are planned for the public launch.

System Architecture

OpenClaw operates as a pipeline of specialized analysis modules, each designed to examine a specific dimension of repository legitimacy. The modules run in parallel where possible and feed their results into a central verdict engine.

Pipeline Overview

// OpenClaw Analysis Pipeline Repository URL |-- Repository Fetcher -- Clone + file tree mapping |-- Code Forensics Engine -- Static analysis per file |-- Commit Analyzer -- Git history pattern detection |-- Dependency Auditor -- Package manifest validation |-- Test Scanner -- Test infrastructure detection |-- Metric Verifier -- README claims vs code reality |-- AI Model Validator -- ML-specific checks | v Verdict Engine -- Confidence-weighted final assessment | v JSON Report + Verdict

Core Components

  • Repository Fetcher — Handles cloning, caching, and file tree mapping. Supports public GitHub repositories with private repo support planned. Generates a complete file manifest with metadata (size, type, last modified).
  • Code Forensics Engine — Performs static analysis on every source file. Checks for function definitions matching claimed functionality, import/usage consistency, execution path coherence, code complexity metrics, and copy-paste detection.
  • Commit Analyzer — Examines the full git history for anomalous patterns. Detects single-day dumps, artificial inflation, bot-like timing, inconsistent authorship, and force-push evidence.
  • Dependency Auditor — Validates package manifests (package.json, requirements.txt, Cargo.toml, etc.) against actual import statements. Detects phantom dependencies, missing dependencies, and version inconsistencies.
  • Test Scanner — Searches for test files, test directories, CI/CD configuration, coverage reports, and test-to-code ratios. Distinguishes between meaningful tests and placeholder stubs.
  • Metric Verifier — Extracts numerical claims from README and documentation, then searches the codebase for evaluation scripts, benchmark datasets, and measurement code that would produce those numbers.
  • AI Model Validator — Specialized module for ML repositories. Checks architecture definitions, model weight availability, training script functionality, and cross-references claims with arXiv, HuggingFace, and PapersWithCode.
  • Verdict Engine — Aggregates signals from all modules using a weighted scoring algorithm. Produces a final verdict (LEGIT or FAKE REPO) with a confidence percentage and detailed evidence breakdown.

Technology Stack

  • AI backbone: Claude (Anthropic) for natural language analysis and reasoning
  • Static analysis: Custom AST parsers for Python, JavaScript, TypeScript, Rust, Go
  • Git analysis: libgit2 bindings for efficient history traversal
  • Infrastructure: Sandboxed execution environments for safe code analysis
  • API: RESTful JSON API with WebSocket support for real-time scan progress

Analysis Methodology

OpenClaw employs a multi-dimensional analysis approach. Each dimension contributes weighted signals to the final verdict, ensuring no single factor produces a false positive or negative.

Code Forensics

Every source file in the repository is analyzed for:

  • Function/class definitions that match README-claimed functionality
  • Import statements that correspond to actual usage in code
  • Execution paths that are logically coherent and reachable
  • Code complexity appropriate for the claimed project scope
  • Dead code ratio — high ratios suggest generated or copy-pasted code
  • Naming conventions and style consistency across the codebase

Commit Pattern Analysis

The git history reveals crucial information about how a project was actually developed. Suspicious patterns include:

  • Single-day dumps where the entire project appears in one or two commits
  • Artificial inflation through meaningless whitespace or formatting changes
  • Bot-like commit timing with unnaturally regular intervals
  • Inconsistent author information suggesting copied histories
  • Evidence of force-pushes that may hide the actual development timeline
  • Commit message patterns that suggest automated or bulk generation

Test Coverage Analysis

Legitimate software projects have testing infrastructure. We scan for:

  • Test files and test directories (tests/, __tests__/, spec/, etc.)
  • CI/CD configuration (.github/workflows, .gitlab-ci.yml, Jenkinsfile)
  • Coverage report artifacts and configuration
  • Test-to-code ratio analysis
  • Quality assessment: meaningful tests vs. placeholder stubs

Metric Verification

When repositories claim specific performance metrics (accuracy, speed, benchmarks), we verify:

  • Presence of evaluation scripts that could produce the claimed numbers
  • Benchmark datasets referenced and accessible
  • Reproducibility of claimed results given the code structure
  • Cross-referencing with published papers when cited

Dependency Audit

  • All declared dependencies are real, published packages
  • Declared dependencies are actually imported in source code
  • No critical dependencies are missing from the manifest
  • Version pinning follows security best practices
  • No known vulnerabilities in dependency tree

Scoring System

The verdict engine uses a weighted multi-signal scoring algorithm to produce two primary metrics: the Legitimacy Score and the Confidence Level.

Legitimacy Score (0-100)

RangeClassificationDescription
90-100Verified LegitimateWell-structured, tested, documented, active development
70-89Likely LegitimateMinor concerns but generally solid codebase
50-69QuestionableSignificant gaps or concerns requiring manual review
30-49SuspiciousMultiple red flags detected across analysis dimensions
0-29Likely FakeOverwhelming evidence of fraud or performative code

Confidence Level

LevelThresholdMeaning
Very High>90%Multiple strong signals align across all dimensions
High70-90%Clear patterns with minor ambiguity
Medium50-70%Mixed signals requiring interpretation
Low<50%Insufficient data for confident assessment

Signal Weights

Each analysis module contributes a weighted signal to the final score. Default weights (configurable via strictness parameter):

code_forensics: 0.30 // Primary signal commit_analysis: 0.20 // History patterns test_coverage: 0.20 // Test infrastructure metric_verification: 0.15 // Claims vs reality dependency_audit: 0.10 // Package validation ai_model_check: 0.05 // ML-specific (if applicable)

API Reference

The ClawGit API provides programmatic access to the scanning engine. Currently in closed beta with limited access.

Scan Repository

POST /api/v1/scan // Request Body { "repo_url": "https://github.com/user/repo", "strictness": 50, "options": { "deep_scan": true, "include_ai_validation": true, "export_format": "json" } }
// Response { "verdict": "FAKE REPO", "confidence": 94, "legitimacy_score": 18, "summary": "All README, no substance", "flags": [ { "code": "ZERO_TESTS", "severity": "critical" }, { "code": "COMMIT_DUMP", "severity": "high" } ], "green_flags": [], "scan_time_ms": 11847, "engine_version": "0.1.0-beta" }

Get Scan Status

GET /api/v1/scan/{scan_id}/status // Response { "status": "in_progress", "progress": 67, "current_module": "commit_analysis", "elapsed_ms": 7200 }

Export Report

GET /api/v1/scan/{scan_id}/report?format=pdf // Returns downloadable PDF report // Status: Coming Soon

Rate Limits

TierLimitStatus
Beta10 scans/hourActive
Pro100 scans/hourPlanned
EnterpriseUnlimitedPlanned

Flag Taxonomy

OpenClaw uses a standardized taxonomy of red and green flags to categorize findings. Each flag is backed by specific evidence from the analysis.

Red Flags

CodeSeverityDescription
GHOST_CODECriticalREADME describes functionality that doesn't exist in code
COMMIT_DUMPHighEntire project committed in a single day
ZERO_TESTSHighNo test files, test configs, or coverage reports
PHANTOM_METRICSHighClaimed performance with no evaluation code
DEAD_DEPSMediumDependencies declared but never imported
STAR_INFLATIONMediumSuspicious star/fork growth patterns
BOILERPLATEMediumMostly template or generated code
MISSING_WEIGHTSHighML repo with no accessible model weights
PAPER_MISMATCHCriticalCode architecture doesn't match claimed paper
COPY_PASTEMediumLarge portions copied from other repositories

Green Flags

CodeSignalDescription
TESTEDStrongComprehensive test suite with meaningful coverage
CI_CDStrongActive continuous integration pipeline
COMMUNITYModerateMultiple contributors, issues, pull requests
DOCUMENTEDModerateInline docs, API docs, usage examples
VERSIONEDModerateProper semantic versioning and changelogs
REPRODUCIBLEStrongClear setup instructions that actually work
BENCHMARKEDStrongVerifiable performance benchmarks included

Integration Guide

ClawGit can be integrated into your development workflow through multiple channels. All integrations use the same underlying OpenClaw engine.

GitHub Actions

# .github/workflows/clawgit.yml name: ClawGit Audit on: [pull_request] jobs: scan: runs-on: ubuntu-latest steps: - uses: clawgit/scan-action@v1 with: strictness: 60 fail-on-verdict: "FAKE REPO" api-key: ${{ secrets.CLAWGIT_API_KEY }}

CLI Tool (Planned)

# Install $ npm install -g @clawgit/cli # Scan a repository $ clawgit scan https://github.com/user/repo # Scan with custom strictness $ clawgit scan https://github.com/user/repo --strictness 80 # Export report $ clawgit scan https://github.com/user/repo --export pdf

Webhook Notifications

// Configure webhook endpoint POST /api/v1/webhooks { "url": "https://your-server.com/clawgit-hook", "events": ["scan.complete", "scan.failed"], "secret": "your-webhook-secret" }

Privacy and Security

  • Repository data is not stored after analysis completes
  • All scan results are encrypted in transit (TLS 1.3)
  • No source code is copied, retained, or used for training
  • Analysis runs in isolated, sandboxed environments
  • API keys are scoped and revocable
  • SOC 2 Type II compliance planned for enterprise launch

Ready to Cut the Larp?

JOIN THE WAITLIST

ClawGit is in closed beta. Be first to access the full OpenClaw engine when we open the gates.

Currently in closed beta -- full launch coming soon