dataset_info:
features:
- name: object_id
dtype: int64
- name: RGB_image
dtype: image
- name: VIS_image
dtype:
array2_d:
shape:
- 224
- 224
dtype: float32
- name: NISP_Y_image
dtype:
array2_d:
shape:
- 224
- 224
dtype: float32
- name: NISP_J_image
dtype:
array2_d:
shape:
- 224
- 224
dtype: float32
- name: NISP_H_image
dtype:
array2_d:
shape:
- 224
- 224
dtype: float32
- name: SED
sequence: float32
splits:
- name: train_batch_1
num_bytes: 3230775343.5
num_examples: 3500
- name: train_batch_10
num_bytes: 3231354520.5
num_examples: 3500
- name: train_batch_11
num_bytes: 3231303636.5
num_examples: 3500
- name: train_batch_12
num_bytes: 3230693275.5
num_examples: 3500
- name: train_batch_13
num_bytes: 3230901080.5
num_examples: 3500
- name: train_batch_14
num_bytes: 3230574418.5
num_examples: 3500
- name: train_batch_15
num_bytes: 3230475865.5
num_examples: 3500
- name: train_batch_16
num_bytes: 3230760051.5
num_examples: 3500
- name: train_batch_17
num_bytes: 3230677954.5
num_examples: 3500
- name: train_batch_18
num_bytes: 3230842166.5
num_examples: 3500
- name: train_batch_19
num_bytes: 3230699299.5
num_examples: 3500
- name: train_batch_2
num_bytes: 3231179908.5
num_examples: 3500
- name: train_batch_20
num_bytes: 3230308880.5
num_examples: 3500
- name: train_batch_21
num_bytes: 3231086424.5
num_examples: 3500
- name: train_batch_22
num_bytes: 3231745385.5
num_examples: 3500
- name: train_batch_23
num_bytes: 3230405393.5
num_examples: 3500
- name: train_batch_24
num_bytes: 3231054305.5
num_examples: 3500
- name: train_batch_25
num_bytes: 3230869946.5
num_examples: 3500
- name: train_batch_26
num_bytes: 3230797298.5
num_examples: 3500
- name: train_batch_27
num_bytes: 3230545728.5
num_examples: 3500
- name: train_batch_28
num_bytes: 3230951549.5
num_examples: 3500
- name: train_batch_29
num_bytes: 3230589123.5
num_examples: 3500
- name: train_batch_3
num_bytes: 3231275355.5
num_examples: 3500
- name: train_batch_30
num_bytes: 3231393228.5
num_examples: 3500
- name: train_batch_31
num_bytes: 3231046330.5
num_examples: 3500
- name: train_batch_32
num_bytes: 3230503816.5
num_examples: 3500
- name: train_batch_33
num_bytes: 3230579197.5
num_examples: 3500
- name: train_batch_34
num_bytes: 3230509280.5
num_examples: 3500
- name: train_batch_35
num_bytes: 3230902018.5
num_examples: 3500
- name: train_batch_36
num_bytes: 3230968470.5
num_examples: 3500
- name: train_batch_37
num_bytes: 3230316762.5
num_examples: 3500
- name: train_batch_38
num_bytes: 3230741330.5
num_examples: 3500
- name: train_batch_39
num_bytes: 3230435587.5
num_examples: 3500
- name: train_batch_4
num_bytes: 3231576444.5
num_examples: 3500
- name: train_batch_40
num_bytes: 3231081419.5
num_examples: 3500
- name: train_batch_41
num_bytes: 3230613876.5
num_examples: 3500
- name: train_batch_42
num_bytes: 3231150193.5
num_examples: 3500
- name: train_batch_43
num_bytes: 3230961683.5
num_examples: 3500
- name: train_batch_44
num_bytes: 3230406514.5
num_examples: 3500
- name: train_batch_45
num_bytes: 3230824323.5
num_examples: 3500
- name: train_batch_46
num_bytes: 3230861870.5
num_examples: 3500
- name: train_batch_47
num_bytes: 3231025981.5
num_examples: 3500
- name: train_batch_48
num_bytes: 3230719191.5
num_examples: 3500
- name: train_batch_49
num_bytes: 3230934872.5
num_examples: 3500
- name: train_batch_5
num_bytes: 3230182077.5
num_examples: 3500
- name: train_batch_50
num_bytes: 3230414329.5
num_examples: 3500
- name: train_batch_51
num_bytes: 3230719288.5
num_examples: 3500
- name: train_batch_52
num_bytes: 3231380990.5
num_examples: 3500
- name: train_batch_53
num_bytes: 3231215638.5
num_examples: 3500
- name: train_batch_54
num_bytes: 3230964153.5
num_examples: 3500
- name: train_batch_55
num_bytes: 3230198581.5
num_examples: 3500
- name: train_batch_56
num_bytes: 3230855019.5
num_examples: 3500
- name: train_batch_57
num_bytes: 3230667706.5
num_examples: 3500
- name: train_batch_58
num_bytes: 3230640904.5
num_examples: 3500
- name: train_batch_6
num_bytes: 3231015571.5
num_examples: 3500
- name: train_batch_60
num_bytes: 3230789065.5
num_examples: 3500
- name: train_batch_61
num_bytes: 3229925818.5
num_examples: 3500
- name: train_batch_62
num_bytes: 3230636974.5
num_examples: 3500
- name: train_batch_63
num_bytes: 3230834456.5
num_examples: 3500
- name: train_batch_64
num_bytes: 3230796391.5
num_examples: 3500
- name: train_batch_65
num_bytes: 3230163774.5
num_examples: 3500
- name: train_batch_66
num_bytes: 3230586797.5
num_examples: 3500
- name: train_batch_67
num_bytes: 3231105894.5
num_examples: 3500
- name: train_batch_68
num_bytes: 3230634477.5
num_examples: 3500
- name: train_batch_69
num_bytes: 3230459406.5
num_examples: 3500
- name: train_batch_7
num_bytes: 3230742078.5
num_examples: 3500
- name: train_batch_70
num_bytes: 3230789645.5
num_examples: 3500
- name: train_batch_71
num_bytes: 3231108986.5
num_examples: 3500
- name: train_batch_72
num_bytes: 3230994836.5
num_examples: 3500
- name: train_batch_73
num_bytes: 3230936220.5
num_examples: 3500
- name: train_batch_74
num_bytes: 3230687237.5
num_examples: 3500
- name: train_batch_75
num_bytes: 3230374591.5
num_examples: 3500
- name: train_batch_76
num_bytes: 2683291658.625
num_examples: 2907
- name: train_batch_8
num_bytes: 3230523632.5
num_examples: 3500
- name: train_batch_9
num_bytes: 3231571589.5
num_examples: 3500
- name: test_batch_1
num_bytes: 3231137560.5
num_examples: 3500
- name: test_batch_10
num_bytes: 3231109888.5
num_examples: 3500
- name: test_batch_11
num_bytes: 3230805519.5
num_examples: 3500
- name: test_batch_12
num_bytes: 3230538042.5
num_examples: 3500
- name: test_batch_13
num_bytes: 3230274570.5
num_examples: 3500
- name: test_batch_14
num_bytes: 3231210165.5
num_examples: 3500
- name: test_batch_15
num_bytes: 3230709431.5
num_examples: 3500
- name: test_batch_16
num_bytes: 3231117584.5
num_examples: 3500
- name: test_batch_17
num_bytes: 3230482570.5
num_examples: 3500
- name: test_batch_18
num_bytes: 3230577234.5
num_examples: 3500
- name: test_batch_19
num_bytes: 3094158486
num_examples: 3352
- name: test_batch_2
num_bytes: 3230956504.5
num_examples: 3500
- name: test_batch_3
num_bytes: 3230104784.5
num_examples: 3500
- name: test_batch_4
num_bytes: 3231582587.5
num_examples: 3500
- name: test_batch_5
num_bytes: 3231120035.5
num_examples: 3500
- name: test_batch_6
num_bytes: 3230450947.5
num_examples: 3500
- name: test_batch_7
num_bytes: 3230849911.5
num_examples: 3500
- name: test_batch_8
num_bytes: 3230900653.5
num_examples: 3500
- name: test_batch_9
num_bytes: 3230604535.5
num_examples: 3500
- name: train_batch_59
num_bytes: 3230954009.5
num_examples: 3500
download_size: 237934118278
dataset_size: 306242272126.125
configs:
- config_name: default
data_files:
- split: train_batch_1
path: data/train_batch_1-*
- split: train_batch_10
path: data/train_batch_10-*
- split: train_batch_11
path: data/train_batch_11-*
- split: train_batch_12
path: data/train_batch_12-*
- split: train_batch_13
path: data/train_batch_13-*
- split: train_batch_14
path: data/train_batch_14-*
- split: train_batch_15
path: data/train_batch_15-*
- split: train_batch_16
path: data/train_batch_16-*
- split: train_batch_17
path: data/train_batch_17-*
- split: train_batch_18
path: data/train_batch_18-*
- split: train_batch_19
path: data/train_batch_19-*
- split: train_batch_2
path: data/train_batch_2-*
- split: train_batch_20
path: data/train_batch_20-*
- split: train_batch_21
path: data/train_batch_21-*
- split: train_batch_22
path: data/train_batch_22-*
- split: train_batch_23
path: data/train_batch_23-*
- split: train_batch_24
path: data/train_batch_24-*
- split: train_batch_25
path: data/train_batch_25-*
- split: train_batch_26
path: data/train_batch_26-*
- split: train_batch_27
path: data/train_batch_27-*
- split: train_batch_28
path: data/train_batch_28-*
- split: train_batch_29
path: data/train_batch_29-*
- split: train_batch_3
path: data/train_batch_3-*
- split: train_batch_30
path: data/train_batch_30-*
- split: train_batch_31
path: data/train_batch_31-*
- split: train_batch_32
path: data/train_batch_32-*
- split: train_batch_33
path: data/train_batch_33-*
- split: train_batch_34
path: data/train_batch_34-*
- split: train_batch_35
path: data/train_batch_35-*
- split: train_batch_36
path: data/train_batch_36-*
- split: train_batch_37
path: data/train_batch_37-*
- split: train_batch_38
path: data/train_batch_38-*
- split: train_batch_39
path: data/train_batch_39-*
- split: train_batch_4
path: data/train_batch_4-*
- split: train_batch_40
path: data/train_batch_40-*
- split: train_batch_41
path: data/train_batch_41-*
- split: train_batch_42
path: data/train_batch_42-*
- split: train_batch_43
path: data/train_batch_43-*
- split: train_batch_44
path: data/train_batch_44-*
- split: train_batch_45
path: data/train_batch_45-*
- split: train_batch_46
path: data/train_batch_46-*
- split: train_batch_47
path: data/train_batch_47-*
- split: train_batch_48
path: data/train_batch_48-*
- split: train_batch_49
path: data/train_batch_49-*
- split: train_batch_5
path: data/train_batch_5-*
- split: train_batch_50
path: data/train_batch_50-*
- split: train_batch_51
path: data/train_batch_51-*
- split: train_batch_52
path: data/train_batch_52-*
- split: train_batch_53
path: data/train_batch_53-*
- split: train_batch_54
path: data/train_batch_54-*
- split: train_batch_55
path: data/train_batch_55-*
- split: train_batch_56
path: data/train_batch_56-*
- split: train_batch_57
path: data/train_batch_57-*
- split: train_batch_58
path: data/train_batch_58-*
- split: train_batch_59
path: data/train_batch_59-*
- split: train_batch_6
path: data/train_batch_6-*
- split: train_batch_60
path: data/train_batch_60-*
- split: train_batch_61
path: data/train_batch_61-*
- split: train_batch_62
path: data/train_batch_62-*
- split: train_batch_63
path: data/train_batch_63-*
- split: train_batch_64
path: data/train_batch_64-*
- split: train_batch_65
path: data/train_batch_65-*
- split: train_batch_66
path: data/train_batch_66-*
- split: train_batch_67
path: data/train_batch_67-*
- split: train_batch_68
path: data/train_batch_68-*
- split: train_batch_69
path: data/train_batch_69-*
- split: train_batch_7
path: data/train_batch_7-*
- split: train_batch_70
path: data/train_batch_70-*
- split: train_batch_71
path: data/train_batch_71-*
- split: train_batch_72
path: data/train_batch_72-*
- split: train_batch_73
path: data/train_batch_73-*
- split: train_batch_74
path: data/train_batch_74-*
- split: train_batch_75
path: data/train_batch_75-*
- split: train_batch_76
path: data/train_batch_76-*
- split: train_batch_8
path: data/train_batch_8-*
- split: train_batch_9
path: data/train_batch_9-*
- split: test_batch_1
path: data/test_batch_1-*
- split: test_batch_10
path: data/test_batch_10-*
- split: test_batch_11
path: data/test_batch_11-*
- split: test_batch_12
path: data/test_batch_12-*
- split: test_batch_13
path: data/test_batch_13-*
- split: test_batch_14
path: data/test_batch_14-*
- split: test_batch_15
path: data/test_batch_15-*
- split: test_batch_16
path: data/test_batch_16-*
- split: test_batch_17
path: data/test_batch_17-*
- split: test_batch_18
path: data/test_batch_18-*
- split: test_batch_19
path: data/test_batch_19-*
- split: test_batch_2
path: data/test_batch_2-*
- split: test_batch_3
path: data/test_batch_3-*
- split: test_batch_4
path: data/test_batch_4-*
- split: test_batch_5
path: data/test_batch_5-*
- split: test_batch_6
path: data/test_batch_6-*
- split: test_batch_7
path: data/test_batch_7-*
- split: test_batch_8
path: data/test_batch_8-*
- split: test_batch_9
path: data/test_batch_9-*
AstroPT Euclid Dataset
A large-scale multi-band astronomical imaging and photometric dataset from the Euclid Q1 dataset, designed for training deep learning models for astronomical analysis.
Overview
This dataset contains 331,759 Euclid bright and large galaxies (HE<22.5 & segmentation_area>0) with simultaneous observations in visible and near-infrared bands from the Euclid mission, complemented by optical photometry from ground-based surveys. It is specifically curated for the (AstroPT) framework for multi-modal learning tasks in astronomy.
Citation: Euclid Collaboration: Siudek, M et al. 2025 (arXiv:2503.15312)
Data Schema
Images
VIS_image(224×224): Visible-band imaging from Euclid VIS instrument (0.55–0.90 μm)NISP_Y_image(224×224): Near-infrared Y-band from Euclid NISP (1.0–1.2 μm)NISP_J_image(224×224): Near-infrared J-band from Euclid NISP (1.2–1.4 μm)NISP_H_image(224×224): Near-infrared H-band from Euclid NISP (1.4–2.0 μm)RGB_image: RGB composite image
Photometry
SED: Spectral Energy Distribution from optical/NIR photometric filters:- u-band (CFHT MegaCam)
- g-band (Subaru HSC)
- r-band (CFHT MegaCam)
- i-band (Pan-STARRS)
- z-band (Subaru HSC)
- g-band (DECam)
- r-band (DECam)
- i-band (DECam)
- z-band (DECam)
- VIS-band (Euclid)
- Y-band (Euclid NISP)
- J-band (Euclid NISP)
- H-band (Euclid NISP)
Metadata
object_id(int64): Unique astronomical object identifier
Quick Start
Load Dataset (Streaming Mode - Recommended)
from datasets import load_dataset
# Load a single training batch without caching
dataset = load_dataset(
"msiudek/astroPT_euclid_dataset",
split="train_batch_1",
streaming=True
)
# Iterate through samples
for sample in dataset:
object_id = sample['object_id']
vis_image = sample['VIS_image'] # 224×224 array
nisp_y = sample['NISP_Y_image'] # 224×224 array
nisp_j = sample['NISP_J_image'] # 224×224 array
nisp_h = sample['NISP_H_image'] # 224×224 array
sed = sample['SED'] # 13 photometric fluxes
# Your processing logic here
Load Multiple Batches
from datasets import load_dataset
from huggingface_hub import list_repo_files
# Get all available splits
files = list_repo_files("msiudek/astroPT_euclid_dataset", repo_type="dataset")
parquets = [f for f in files if '.parquet' in f]
batches = set(f.split('/')[-1].rsplit('-', 2)[0] for f in parquets)
# Process all training batches
train_batches = sorted([b for b in batches if 'train_batch' in b])
for batch_name in train_batches:
try:
dataset = load_dataset(
"msiudek/astroPT_euclid_dataset",
split=batch_name,
streaming=True
)
for sample in dataset:
# Process sample
pass
except Exception as e:
print(f"Error loading {batch_name}: {e}")
Load Non-Streaming (Download to Disk)
dataset = load_dataset(
"msiudek/astroPT_euclid_dataset",
streaming=False,
cache_dir="/path/to/cache"
)
Data Access & Citation
Framework: AstroPT GitHub Repository
- Multi-modal learning framework
- Pre-training and fine-tuning code
- Example notebooks and tutorials
Metadata: astroPT_euclid_metadata
- Morphological properties from Euclid
- Physical properties (stellar mass, SFR, etc.)
- Spectroscopic redshifts from DESI (~5% of sample)
Citation:
@article{Siudek2025,
title={AstroPT: Astronomical Physics Transformers for Multi-modal Learning},
author={Siudek, M and others},
journal={Euclid Collaboration},
eprint={2503.15312},
archivePrefix={arXiv},
year={2025},
url={https://ui.adsabs.harvard.edu/abs/2025arXiv250315312E/abstract}
}
Last Updated: December 2025
Dataset Version: 1.0
Euclid Release: Q1
Data License: CC-BY-4.0
Acknowledgments
This dataset was generated using data from the Euclid mission, an ESA medium-class astronomy and astrophysics space mission. Euclid is a space mission of the European Space Agency (ESA).