openllm / README.md
lemms's picture
Update README for Gradio 4.44.1 and complete training features
d185e19 verified
|
raw
history blame
2.81 kB
metadata
title: OpenLLM Training Space
emoji: πŸš€
colorFrom: blue
colorTo: purple
sdk: gradio
sdk_version: 4.44.1
app_file: app.py
pinned: false
license: gpl-3.0

OpenLLM Training Space

This space provides complete training infrastructure for OpenLLM models with real model training functionality.

Features

  • 🎯 Real Model Training: Actual PyTorch training with Transformers
  • πŸ“Š Training Monitoring: Live progress tracking and loss monitoring
  • πŸ”„ Model Versioning: Automatic model saving and uploading to HF Hub
  • πŸ“ˆ Performance Tracking: Training metrics and completion status
  • πŸš€ Gradio 4.44.1: Latest UI framework with enhanced compatibility

Complete Training Pipeline

What Happens When You Click "Start Training":

  1. πŸ“₯ Model Loading: Loads the 7k model from lemms/openllm-small-extended-7k
  2. πŸ“Š Dataset Preparation: Loads and tokenizes training data from lemms/openllm-training-data
  3. βš™οΈ Training Setup: Configures PyTorch Trainer with your parameters
  4. πŸš€ Real Training: Executes actual model training for specified steps
  5. πŸ’Ύ Save & Upload: Saves trained model and uploads to HF Hub as lemms/openllm-{size}-extended-8k

Training Configuration Options:

  • Model Size: small, medium, large (currently supports small)
  • Max Steps: 100-10,000 training iterations
  • Learning Rate: 0.00001-0.001 (configurable)
  • Batch Size: 1-16 samples per batch

Expected Results:

  • Training Time: 10-30 minutes for 1000 steps (depending on HF Space resources)
  • Output Model: lemms/openllm-small-extended-8k (or other sizes)
  • Model Files: Complete PyTorch model with tokenizer and configuration

Model Repositories

Technical Details

  • Framework: PyTorch with Transformers
  • UI: Gradio 4.44.1 (latest stable version)
  • Training: Mixed precision (FP16) for efficiency
  • Memory: Optimized for HF Spaces with gradient accumulation
  • Dependencies: Complete ML stack with all training utilities

Usage

  1. Configure Parameters: Set model size, steps, learning rate, and batch size
  2. Start Training: Click "Start Training" to begin the complete pipeline
  3. Monitor Progress: Watch real-time status updates and training progress
  4. Access Results: Find your trained model in the HF Hub repository

License

GPL-3.0 - See LICENSE for details.

Author

Louis Chua Bean Chong - OpenLLM Project Maintainer