Dataset Viewer
Auto-converted to Parquet Duplicate
experiment_id
stringlengths
19
24
name
stringlengths
18
45
description
stringlengths
20
71
created_at
stringdate
2025-07-20 11:20:01
2025-08-09 12:24:12
status
stringclasses
2 values
metrics
stringlengths
2
163k
parameters
stringlengths
2
4.09k
artifacts
stringclasses
2 values
logs
stringclasses
5 values
last_updated
stringdate
2025-08-09 17:43:05
2025-08-09 17:43:05
exp_20250720_130853
petite-elle-l-aime-3
SmolLM3 fine-tuning experiment
2025-07-20T11:20:01.780908
running
[{"timestamp": "2025-07-20T11:20:01.780908", "step": 25, "metrics": {"loss": 1.1659, "grad_norm": 10.3125, "learning_rate": 7e-08, "num_tokens": 1642080.0, "mean_token_accuracy": 0.75923578992486, "epoch": 0.004851130919895701}}, {"timestamp": "2025-07-20T11:26:39.042155", "step": 50, "metrics": {"loss": 1.165, "grad_n...
{"model_name": "HuggingFaceTB/SmolLM3-3B", "max_seq_length": 12288, "use_flash_attention": true, "use_gradient_checkpointing": false, "batch_size": 8, "gradient_accumulation_steps": 16, "learning_rate": 3.5e-06, "weight_decay": 0.01, "warmup_steps": 1200, "max_iters": 18000, "eval_interval": 1000, "log_interval": 25, "...
[]
[]
2025-08-09T17:43:05.368860
exp_20250720_134319
petite-elle-l-aime-3-1
SmolLM3 fine-tuning experiment
2025-07-20T11:54:31.993219
running
[{"timestamp": "2025-07-20T11:54:33.589487", "step": 25, "metrics": {"loss": 1.166, "grad_norm": 10.375, "learning_rate": 7e-08, "num_tokens": 1642080.0, "mean_token_accuracy": 0.7590958896279335, "epoch": 0.004851130919895701, "gpu_0_memory_allocated": 17.202261447906494, "gpu_0_memory_reserved": 75.474609375, "gpu_0_...
{"model_name": "HuggingFaceTB/SmolLM3-3B", "max_seq_length": 12288, "use_flash_attention": true, "use_gradient_checkpointing": false, "batch_size": 8, "gradient_accumulation_steps": 16, "learning_rate": 3.5e-06, "weight_decay": 0.01, "warmup_steps": 1200, "max_iters": 18000, "eval_interval": 1000, "log_interval": 25, "...
[]
[]
2025-08-09T17:43:05.369029
exp_20250727_172507
petite_llm_3_fr_1_20250727_152506
SmolLM3 fine-tuning experiment: petite_llm_3_fr_1
2025-07-27T17:25:07.131302
running
[]
{"trainer_type": "sft", "model_name": "HuggingFaceTB/SmolLM3-3B", "max_seq_length": 8192, "use_flash_attention": true, "use_gradient_checkpointing": false, "batch_size": 8, "gradient_accumulation_steps": 16, "learning_rate": 5e-06, "weight_decay": 0.01, "warmup_steps": 1000, "max_iters": 8000, "eval_interval": 500, "lo...
[]
[]
2025-08-09T17:43:05.369139
exp_20250727_172526
petite_llm_3_fr_1_20250727_152525
SmolLM3 fine-tuning experiment: petite_llm_3_fr_1
2025-07-27T17:25:26.109242
completed
[{"timestamp": "2025-07-27T17:37:01.100450", "step": 25, "metrics": {"loss": 1.1733, "grad_norm": 11.25, "learning_rate": 1.2000000000000002e-07, "num_tokens": 1642080.0, "mean_token_accuracy": 0.7592912124097347, "epoch": 0.004851130919895701, "timestamp": "2025-07-27T15:37:00.527604", "step": 25, "gpu_0_memory_alloca...
{"trainer_type": "sft", "model_name": "HuggingFaceTB/SmolLM3-3B", "max_seq_length": 8192, "use_flash_attention": true, "use_gradient_checkpointing": false, "batch_size": 8, "gradient_accumulation_steps": 16, "learning_rate": 5e-06, "weight_decay": 0.01, "warmup_steps": 1000, "max_iters": 8000, "eval_interval": 500, "lo...
[]
[]
2025-08-09T17:43:05.381795
exp_20250727_172538
smollm3_experiment_20250727_152538
SmolLM3 fine-tuning experiment: smollm3_experiment
2025-07-27T17:25:38.978779
running
[]
{}
[]
[]
2025-08-09T17:43:05.382137
exp_20250727_182356
test_diagnosis_20250727_182356
Diagnosis test experiment
2025-07-27T18:23:56.924122
running
[{"timestamp": "2025-07-27T18:24:07.872673", "step": 100, "metrics": {"loss": 1.234, "accuracy": 0.85}}]
{}
[]
[]
2025-08-09T17:43:05.382250
exp_20250727_182415
test_monitoring_diagnosis_20250727_182414
SmolLM3 fine-tuning experiment: test_monitoring_diagnosis
2025-07-27T18:24:15.047914
running
[{"timestamp": "2025-07-27T18:24:19.301319", "step": 200, "metrics": {"loss": 2.345, "accuracy": 0.75, "timestamp": "2025-07-27T18:24:18.657441", "step": 200}}]
{"learning_rate": 2e-05, "batch_size": 8}
[]
[]
2025-08-09T17:43:05.382352
exp_20250727_182446
flow_test_20250727_182445
Flow test experiment
2025-07-27T18:24:46.577375
running
[]
{}
[]
[]
2025-08-09T17:43:05.382420
exp_20250727_182703
validation_test_20250727_182703
Validation test experiment
2025-07-27T18:27:03.714216
running
[]
{}
[]
[]
2025-08-09T17:43:05.382486
exp_20250727_182718
test_recreation_20250727_182717
SmolLM3 fine-tuning experiment: test_recreation
2025-07-27T18:27:18.185932
running
[]
{}
[]
[]
2025-08-09T17:43:05.382550
exp_20250727_182734
test_recreation_recreated_20250727_182733
Recreated SmolLM3 fine-tuning experiment: test_recreation
2025-07-27T18:27:34.321157
running
[]
{"experiment_recreated": true, "original_experiment_name": "test_recreation", "recreation_timestamp": "20250727_182733", "recreation_reason": "Original experiment not found or expired"}
[]
[]
2025-08-09T17:43:05.382624
exp_20250727_182744
test_recreation_recreated_20250727_182744
Recreated SmolLM3 fine-tuning experiment: test_recreation
2025-07-27T18:27:44.894014
running
[]
{"experiment_recreated": true, "original_experiment_name": "test_recreation", "recreation_timestamp": "20250727_182744", "recreation_reason": "Original experiment not found or expired"}
[]
[]
2025-08-09T17:43:05.382697
exp_20250727_182802
test_robust_logging_20250727_182801
SmolLM3 fine-tuning experiment: test_robust_logging
2025-07-27T18:28:02.317093
running
[{"timestamp": "2025-07-27T18:28:15.147475", "step": 100, "metrics": {"loss": 1.5, "accuracy": 0.8, "timestamp": "2025-07-27T18:28:11.283323", "step": 100}}]
{"learning_rate": 2e-05, "batch_size": 8}
[]
[]
2025-08-09T17:43:05.382788
exp_20250727_182824
test_robust_logging_recreated_20250727_182824
Recreated SmolLM3 fine-tuning experiment: test_robust_logging
2025-07-27T18:28:24.600453
running
[]
{"experiment_recreated": true, "original_experiment_name": "test_robust_logging", "recreation_timestamp": "20250727_182824", "recreation_reason": "Original experiment not found or expired"}
[]
[]
2025-08-09T17:43:05.382863
exp_20250727_182833
test_robust_logging_recreated_20250727_182833
Recreated SmolLM3 fine-tuning experiment: test_robust_logging
2025-07-27T18:28:33.707029
running
[]
{"experiment_recreated": true, "original_experiment_name": "test_robust_logging", "recreation_timestamp": "20250727_182833", "recreation_reason": "Original experiment not found or expired"}
[]
[]
2025-08-09T17:43:05.382935
exp_20250727_183001
simple_test_20250727_183000
SmolLM3 fine-tuning experiment: simple_test
2025-07-27T18:30:01.254197
running
[{"timestamp": "2025-07-27T18:30:08.455622", "step": 1, "metrics": {"loss": 1.0, "accuracy": 0.8, "timestamp": "2025-07-27T18:30:05.410525", "step": 1}}]
{"learning_rate": 2e-05, "batch_size": 8}
[]
[]
2025-08-09T17:43:05.383024
exp_20250727_193248
continuity_test_20250727_193248
SmolLM3 fine-tuning experiment: continuity_test
2025-07-27T19:32:48.780338
running
[{"timestamp": "2025-07-27T19:32:55.332253", "step": 0, "metrics": {"loss": 2.5, "learning_rate": 0.0001, "step": 0, "phase": "initial", "timestamp": "2025-07-27T19:32:52.440576"}}, {"timestamp": "2025-07-27T19:33:01.142884", "step": 10, "metrics": {"loss": 2.3, "learning_rate": 0.0001, "step": 10, "phase": "post_loss"...
{"model_name": "SmolLM3-3B", "dataset": "OpenHermes-FR", "batch_size": 8, "learning_rate": 0.0001, "max_steps": 1000, "continuity_test": true}
[]
[]
2025-08-09T17:43:05.383195
exp_20250727_193347
multiple_recreations_test_20250727_193347
SmolLM3 fine-tuning experiment: multiple_recreations_test
2025-07-27T19:33:47.807935
running
[{"timestamp": "2025-07-27T19:33:55.433909", "step": 0, "metrics": {"loss": 2.0, "step": 0, "recreation_count": 1, "timestamp": "2025-07-27T19:33:52.457912"}}, {"timestamp": "2025-07-27T19:34:01.265497", "step": 10, "metrics": {"loss": 1.8, "step": 10, "recreation_count": 2, "timestamp": "2025-07-27T19:33:58.178631"}},...
{}
[]
[]
2025-08-09T17:43:05.383375
exp_demo_20250808_154602
smollm3-finetune-demo
SmolLM3 fine-tuning experiment demo with comprehensive metrics tracking
2025-08-08T15:46:02.531457
completed
[{"timestamp": "2025-08-08T15:46:02.531462", "step": 100, "metrics": {"loss": 1.15, "grad_norm": 10.5, "learning_rate": 5e-06, "num_tokens": 1000000.0, "mean_token_accuracy": 0.76, "epoch": 0.1, "total_tokens": 1000000.0, "throughput": 2000000.0, "step_time": 0.5, "batch_size": 2, "seq_len": 4096, "token_acc": 0.76, "g...
{"model_name": "HuggingFaceTB/SmolLM3-3B", "max_seq_length": 4096, "batch_size": 2, "learning_rate": 5e-06, "epochs": 3, "dataset": "OpenHermes-FR", "trainer_type": "SFTTrainer", "hardware": "GPU (H100/A100)", "mixed_precision": true, "gradient_checkpointing": true, "flash_attention": true}
[]
[{"timestamp": "2025-08-08T15:46:02.531537", "level": "INFO", "message": "Training started successfully"}, {"timestamp": "2025-08-08T15:46:02.531542", "level": "INFO", "message": "Model loaded and configured"}, {"timestamp": "2025-08-08T15:46:02.531545", "level": "INFO", "message": "Dataset loaded and preprocessed"}]
2025-08-09T17:43:05.383586
End of preview. Expand in Data Studio

Trackio Experiments Dataset

This dataset stores experiment tracking data for ML training runs, particularly focused on SmolLM3 fine-tuning experiments with comprehensive metrics tracking.

Dataset Structure

The dataset contains the following columns:

  • experiment_id: Unique identifier for each experiment
  • name: Human-readable name for the experiment
  • description: Detailed description of the experiment
  • created_at: Timestamp when the experiment was created
  • status: Current status (running, completed, failed, paused)
  • metrics: JSON string containing training metrics over time
  • parameters: JSON string containing experiment configuration
  • artifacts: JSON string containing experiment artifacts
  • logs: JSON string containing experiment logs
  • last_updated: Timestamp of last update

Metrics Structure

The metrics field contains JSON arrays with the following structure:

[
  {
    "timestamp": "2025-07-20T11:20:01.780908",
    "step": 25,
    "metrics": {
      "loss": 1.1659,
      "accuracy": 0.759,
      "learning_rate": 7e-08,
      "grad_norm": 10.3125,
      "epoch": 0.004851130919895701,
      
      // Advanced Training Metrics
      "total_tokens": 1642080.0,
      "truncated_tokens": 128,
      "padding_tokens": 256,
      "throughput": 3284160.0,
      "step_time": 0.5,
      "batch_size": 8,
      "seq_len": 2048,
      "token_acc": 0.759,
      
      // Custom Losses
      "train/gate_ortho": 0.0234,
      "train/center": 0.0156,
      
      // System Metrics
      "gpu_memory_allocated": 17.202261447906494,
      "gpu_memory_reserved": 75.474609375,
      "gpu_utilization": 85.2,
      "cpu_percent": 2.7,
      "memory_percent": 10.1
    }
  }
]

Supported Metrics

Core Training Metrics

  • loss: Training loss value
  • accuracy: Model accuracy
  • learning_rate: Current learning rate
  • grad_norm: Gradient norm
  • epoch: Current epoch progress

Advanced Token Metrics

  • total_tokens: Total tokens processed in the batch
  • truncated_tokens: Number of tokens truncated during processing
  • padding_tokens: Number of padding tokens added
  • throughput: Tokens processed per second
  • step_time: Time taken for the current training step
  • batch_size: Current batch size
  • seq_len: Sequence length
  • token_acc: Token-level accuracy

Custom Losses (SmolLM3-specific)

  • train/gate_ortho: Gate orthogonality loss
  • train/center: Center loss component

System Performance Metrics

  • gpu_memory_allocated: GPU memory currently allocated (GB)
  • gpu_memory_reserved: GPU memory reserved (GB)
  • gpu_utilization: GPU utilization percentage
  • cpu_percent: CPU usage percentage
  • memory_percent: System memory usage percentage

Usage

This dataset is automatically used by the Trackio monitoring system to store and retrieve experiment data. It provides persistent storage for experiment tracking across different training runs.

Integration

The dataset is used by:

  • Trackio Spaces for experiment visualization
  • Training scripts for logging metrics and parameters
  • Monitoring systems for experiment tracking
  • SmolLM3 fine-tuning pipeline for comprehensive metrics capture

Privacy

This dataset is private by default to ensure experiment data security. Only users with appropriate permissions can access the data.

Examples

Sample Experiment Entry

{
  "experiment_id": "exp_20250720_130853",
  "name": "smollm3_finetune",
  "description": "SmolLM3 fine-tuning experiment with comprehensive metrics",
  "created_at": "2025-07-20T11:20:01.780908",
  "status": "running",
  "metrics": "[{\"timestamp\": \"2025-07-20T11:20:01.780908\", \"step\": 25, \"metrics\": {\"loss\": 1.1659, \"accuracy\": 0.759, \"total_tokens\": 1642080.0, \"throughput\": 3284160.0, \"train/gate_ortho\": 0.0234, \"train/center\": 0.0156}}]",
  "parameters": "{\"model_name\": \"HuggingFaceTB/SmolLM3-3B\", \"batch_size\": 8, \"learning_rate\": 3.5e-06, \"max_seq_length\": 12288}",
  "artifacts": "[]",
  "logs": "[]",
  "last_updated": "2025-07-20T11:20:01.780908"
}

License

This dataset is part of the Trackio experiment tracking system and follows the same license as the main project.

Downloads last month
179

Spaces using Tonic/trackio-experiments 5