Multi-modal Earth Observation pretraining dataset in TACO format. Part A: 250k monotemporal tiles, Part B: ~12.5k locations × 12 monthly timesteps, Part C: ~16.6k locations × 6 five-daily timesteps. Sensors: Sentinel-2 L1C, Landsat 8/9, Copernicus DEM, ESA WorldCover. Locations selected via AEF hierarchical clustering over the MajorTOM 10km grid. Funded by the ELLIOT project.
ELLIOT-Pretrain is a Major TOM expansion focused on fast pre-training of multi-modal AI foundation models on Earth Observation data.
Tile locations were selected using hierarchical spherical k-means clustering over AlphaEarth Foundation embeddings to maximise global environmental diversity. The dataset follows the TACO v3 specification.
All three parts share the same four sensors.
Pick a tile index from Part A (monotemporal) and visualize all four modalities.
Full notebook covering all three parts, metadata queries with filtering, and a streaming PyTorch DataLoader with parallel fetching.
Join key across metadata files is internal:parent_id, which points back to internal:current_id in collection.parquet.
CC-BY-SA-4.0
ELLIOT-Pretrain has been made possible thanks to Asterisk Labs, the ELLIOT project (European Commission, Horizon Europe, Grant 101214398), and the Image and Signal Processing Group (ISP) at Universitat de Valencia.