Skip to content

An interactive dashboard for analyzing and comparing the performance, cost, and capabilities of leading Large Language Models (LLMs). This tool fetches real-time data from the Hugging Face Open LLM Leaderboard and a custom data repository to provide up-to-date insights.

License

Notifications You must be signed in to change notification settings

codebywiam/llm-analytics-dashboard

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

1 Commit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LLM Analytics Dashboard

alt text alt text

An interactive dashboard for analyzing and comparing the performance, cost, and capabilities of leading Large Language Models (LLMs). This tool fetches real-time data from the Hugging Face Open LLM Leaderboard and a custom data repository to provide up-to-date insights.

Features

  • Statistics View:
    • GPQA Score Timeline: Tracks the evolution of model performance (GPQA benchmark) over time, with interactive logos for major AI organizations.
    • Context Window Comparison: A comprehensive bar chart comparing the input and output context lengths of various models.
    • Performance vs. Cost Analysis: A scatter plot identifying the most cost-effective models based on their GPQA score and cost per million tokens.
  • Comparator View:
    • Dynamic Model Selection: Search and select multiple models for a side-by-side comparison.
    • Benchmark Breakdown: Compares selected models across key benchmarks like MMLU-Pro, BBH, and MATH.
    • CO₂ Impact: Visualizes the carbon cost of training, highlighting the environmental impact relative to model size.
    • Detailed Data Table: Provides a sortable and filterable table with raw metrics for in-depth analysis.

Tech Stack

  • Backend & Dashboard: Python, Dash, Plotly
  • Data Processing: Pandas, NumPy
  • Data Sources: Hugging Face datasets API, GitHub API
  • Performance: In-memory caching with time-based invalidation and parallel data fetching using ThreadPoolExecutor.

Setup & Installation

  1. Clone the repository:

    git clone https://github.com/codebywiam/llm-analytics-dashboard.git
    cd llm-analytics-dashboard
  2. Create and activate a virtual environment:

    python -m venv venv
    source venv/bin/activate  # On Windows, use `venv\Scripts\activate`
  3. Install dependencies:

    pip install -r requirements.txt
  4. Run the application:

    python app.py

    The app will be available at http://127.0.0.1:8050.

Dashboard

Statistics

alt text alt text alt text alt text

Comparator

alt text alt text alt text alt text

Future Improvements

  • Implement a more robust data versioning system instead of relying solely on live APIs.
  • Add unit and integration tests for data processing functions.
  • Containerize the application with Docker for easier deployment.

License

This project is licensed under the MIT License. See LICENSE for details.

About

An interactive dashboard for analyzing and comparing the performance, cost, and capabilities of leading Large Language Models (LLMs). This tool fetches real-time data from the Hugging Face Open LLM Leaderboard and a custom data repository to provide up-to-date insights.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published