The dataset viewer is not available for this split.
Parquet error: Scan size limit exceeded: attempted to read 303734161 bytes, limit is 300000000 bytes
Make sure that
1. the Parquet files contain a page index to enable random access without loading entire row groups2. otherwise use smaller row-group sizes when serializing the Parquet files
Error code: TooBigContentError
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
Changelog
NEW Changes July 25th
- added
baseModelsfield to models which shows the models that the user tagged as base models for that model
Example:
{
"models": [
{
"_id": "687de260234339fed21e768a",
"id": "Qwen/Qwen3-235B-A22B-Instruct-2507"
}
],
"relation": "quantized"
}
NEW Changes July 9th
- Fixed issue with
ggufcolumn with integer overflow causing import pipeline to be broken over a few weeks ✅
NEW Changes Feb 27th
Added new fields on the
modelssplit:downloadsAllTime,safetensors,ggufAdded new field on the
datasetssplit:downloadsAllTimeAdded new split:
paperswhich is all of the Daily Papers
Updated Daily
- Downloads last month
- 4,806
Size of downloaded dataset files:
1.79 GB
Size of the auto-converted Parquet files:
1.79 GB
Number of rows:
4,359,170