--- dataset_info: features: - name: object_id dtype: int64 - name: RGB_image dtype: image - name: VIS_image dtype: array2_d: shape: - 224 - 224 dtype: float32 - name: NISP_Y_image dtype: array2_d: shape: - 224 - 224 dtype: float32 - name: NISP_J_image dtype: array2_d: shape: - 224 - 224 dtype: float32 - name: NISP_H_image dtype: array2_d: shape: - 224 - 224 dtype: float32 - name: SED sequence: float32 splits: - name: train_batch_1 num_bytes: 3230775343.5 num_examples: 3500 - name: train_batch_10 num_bytes: 3231354520.5 num_examples: 3500 - name: train_batch_11 num_bytes: 3231303636.5 num_examples: 3500 - name: train_batch_12 num_bytes: 3230693275.5 num_examples: 3500 - name: train_batch_13 num_bytes: 3230901080.5 num_examples: 3500 - name: train_batch_14 num_bytes: 3230574418.5 num_examples: 3500 - name: train_batch_15 num_bytes: 3230475865.5 num_examples: 3500 - name: train_batch_16 num_bytes: 3230760051.5 num_examples: 3500 - name: train_batch_17 num_bytes: 3230677954.5 num_examples: 3500 - name: train_batch_18 num_bytes: 3230842166.5 num_examples: 3500 - name: train_batch_19 num_bytes: 3230699299.5 num_examples: 3500 - name: train_batch_2 num_bytes: 3231179908.5 num_examples: 3500 - name: train_batch_20 num_bytes: 3230308880.5 num_examples: 3500 - name: train_batch_21 num_bytes: 3231086424.5 num_examples: 3500 - name: train_batch_22 num_bytes: 3231745385.5 num_examples: 3500 - name: train_batch_23 num_bytes: 3230405393.5 num_examples: 3500 - name: train_batch_24 num_bytes: 3231054305.5 num_examples: 3500 - name: train_batch_25 num_bytes: 3230869946.5 num_examples: 3500 - name: train_batch_26 num_bytes: 3230797298.5 num_examples: 3500 - name: train_batch_27 num_bytes: 3230545728.5 num_examples: 3500 - name: train_batch_28 num_bytes: 3230951549.5 num_examples: 3500 - name: train_batch_29 num_bytes: 3230589123.5 num_examples: 3500 - name: train_batch_3 num_bytes: 3231275355.5 num_examples: 3500 - name: train_batch_30 num_bytes: 3231393228.5 num_examples: 3500 - name: train_batch_31 num_bytes: 3231046330.5 num_examples: 3500 - name: train_batch_32 num_bytes: 3230503816.5 num_examples: 3500 - name: train_batch_33 num_bytes: 3230579197.5 num_examples: 3500 - name: train_batch_34 num_bytes: 3230509280.5 num_examples: 3500 - name: train_batch_35 num_bytes: 3230902018.5 num_examples: 3500 - name: train_batch_36 num_bytes: 3230968470.5 num_examples: 3500 - name: train_batch_37 num_bytes: 3230316762.5 num_examples: 3500 - name: train_batch_38 num_bytes: 3230741330.5 num_examples: 3500 - name: train_batch_39 num_bytes: 3230435587.5 num_examples: 3500 - name: train_batch_4 num_bytes: 3231576444.5 num_examples: 3500 - name: train_batch_40 num_bytes: 3231081419.5 num_examples: 3500 - name: train_batch_41 num_bytes: 3230613876.5 num_examples: 3500 - name: train_batch_42 num_bytes: 3231150193.5 num_examples: 3500 - name: train_batch_43 num_bytes: 3230961683.5 num_examples: 3500 - name: train_batch_44 num_bytes: 3230406514.5 num_examples: 3500 - name: train_batch_45 num_bytes: 3230824323.5 num_examples: 3500 - name: train_batch_46 num_bytes: 3230861870.5 num_examples: 3500 - name: train_batch_47 num_bytes: 3231025981.5 num_examples: 3500 - name: train_batch_48 num_bytes: 3230719191.5 num_examples: 3500 - name: train_batch_49 num_bytes: 3230934872.5 num_examples: 3500 - name: train_batch_5 num_bytes: 3230182077.5 num_examples: 3500 - name: train_batch_50 num_bytes: 3230414329.5 num_examples: 3500 - name: train_batch_51 num_bytes: 3230719288.5 num_examples: 3500 - name: train_batch_52 num_bytes: 3231380990.5 num_examples: 3500 - name: train_batch_53 num_bytes: 3231215638.5 num_examples: 3500 - name: train_batch_54 num_bytes: 3230964153.5 num_examples: 3500 - name: train_batch_55 num_bytes: 3230198581.5 num_examples: 3500 - name: train_batch_56 num_bytes: 3230855019.5 num_examples: 3500 - name: train_batch_57 num_bytes: 3230667706.5 num_examples: 3500 - name: train_batch_58 num_bytes: 3230640904.5 num_examples: 3500 - name: train_batch_6 num_bytes: 3231015571.5 num_examples: 3500 - name: train_batch_60 num_bytes: 3230789065.5 num_examples: 3500 - name: train_batch_61 num_bytes: 3229925818.5 num_examples: 3500 - name: train_batch_62 num_bytes: 3230636974.5 num_examples: 3500 - name: train_batch_63 num_bytes: 3230834456.5 num_examples: 3500 - name: train_batch_64 num_bytes: 3230796391.5 num_examples: 3500 - name: train_batch_65 num_bytes: 3230163774.5 num_examples: 3500 - name: train_batch_66 num_bytes: 3230586797.5 num_examples: 3500 - name: train_batch_67 num_bytes: 3231105894.5 num_examples: 3500 - name: train_batch_68 num_bytes: 3230634477.5 num_examples: 3500 - name: train_batch_69 num_bytes: 3230459406.5 num_examples: 3500 - name: train_batch_7 num_bytes: 3230742078.5 num_examples: 3500 - name: train_batch_70 num_bytes: 3230789645.5 num_examples: 3500 - name: train_batch_71 num_bytes: 3231108986.5 num_examples: 3500 - name: train_batch_72 num_bytes: 3230994836.5 num_examples: 3500 - name: train_batch_73 num_bytes: 3230936220.5 num_examples: 3500 - name: train_batch_74 num_bytes: 3230687237.5 num_examples: 3500 - name: train_batch_75 num_bytes: 3230374591.5 num_examples: 3500 - name: train_batch_76 num_bytes: 2683291658.625 num_examples: 2907 - name: train_batch_8 num_bytes: 3230523632.5 num_examples: 3500 - name: train_batch_9 num_bytes: 3231571589.5 num_examples: 3500 - name: test_batch_1 num_bytes: 3231137560.5 num_examples: 3500 - name: test_batch_10 num_bytes: 3231109888.5 num_examples: 3500 - name: test_batch_11 num_bytes: 3230805519.5 num_examples: 3500 - name: test_batch_12 num_bytes: 3230538042.5 num_examples: 3500 - name: test_batch_13 num_bytes: 3230274570.5 num_examples: 3500 - name: test_batch_14 num_bytes: 3231210165.5 num_examples: 3500 - name: test_batch_15 num_bytes: 3230709431.5 num_examples: 3500 - name: test_batch_16 num_bytes: 3231117584.5 num_examples: 3500 - name: test_batch_17 num_bytes: 3230482570.5 num_examples: 3500 - name: test_batch_18 num_bytes: 3230577234.5 num_examples: 3500 - name: test_batch_19 num_bytes: 3094158486.0 num_examples: 3352 - name: test_batch_2 num_bytes: 3230956504.5 num_examples: 3500 - name: test_batch_3 num_bytes: 3230104784.5 num_examples: 3500 - name: test_batch_4 num_bytes: 3231582587.5 num_examples: 3500 - name: test_batch_5 num_bytes: 3231120035.5 num_examples: 3500 - name: test_batch_6 num_bytes: 3230450947.5 num_examples: 3500 - name: test_batch_7 num_bytes: 3230849911.5 num_examples: 3500 - name: test_batch_8 num_bytes: 3230900653.5 num_examples: 3500 - name: test_batch_9 num_bytes: 3230604535.5 num_examples: 3500 - name: train_batch_59 num_bytes: 3230954009.5 num_examples: 3500 download_size: 237934118278 dataset_size: 306242272126.125 configs: - config_name: default data_files: - split: train_batch_1 path: data/train_batch_1-* - split: train_batch_10 path: data/train_batch_10-* - split: train_batch_11 path: data/train_batch_11-* - split: train_batch_12 path: data/train_batch_12-* - split: train_batch_13 path: data/train_batch_13-* - split: train_batch_14 path: data/train_batch_14-* - split: train_batch_15 path: data/train_batch_15-* - split: train_batch_16 path: data/train_batch_16-* - split: train_batch_17 path: data/train_batch_17-* - split: train_batch_18 path: data/train_batch_18-* - split: train_batch_19 path: data/train_batch_19-* - split: train_batch_2 path: data/train_batch_2-* - split: train_batch_20 path: data/train_batch_20-* - split: train_batch_21 path: data/train_batch_21-* - split: train_batch_22 path: data/train_batch_22-* - split: train_batch_23 path: data/train_batch_23-* - split: train_batch_24 path: data/train_batch_24-* - split: train_batch_25 path: data/train_batch_25-* - split: train_batch_26 path: data/train_batch_26-* - split: train_batch_27 path: data/train_batch_27-* - split: train_batch_28 path: data/train_batch_28-* - split: train_batch_29 path: data/train_batch_29-* - split: train_batch_3 path: data/train_batch_3-* - split: train_batch_30 path: data/train_batch_30-* - split: train_batch_31 path: data/train_batch_31-* - split: train_batch_32 path: data/train_batch_32-* - split: train_batch_33 path: data/train_batch_33-* - split: train_batch_34 path: data/train_batch_34-* - split: train_batch_35 path: data/train_batch_35-* - split: train_batch_36 path: data/train_batch_36-* - split: train_batch_37 path: data/train_batch_37-* - split: train_batch_38 path: data/train_batch_38-* - split: train_batch_39 path: data/train_batch_39-* - split: train_batch_4 path: data/train_batch_4-* - split: train_batch_40 path: data/train_batch_40-* - split: train_batch_41 path: data/train_batch_41-* - split: train_batch_42 path: data/train_batch_42-* - split: train_batch_43 path: data/train_batch_43-* - split: train_batch_44 path: data/train_batch_44-* - split: train_batch_45 path: data/train_batch_45-* - split: train_batch_46 path: data/train_batch_46-* - split: train_batch_47 path: data/train_batch_47-* - split: train_batch_48 path: data/train_batch_48-* - split: train_batch_49 path: data/train_batch_49-* - split: train_batch_5 path: data/train_batch_5-* - split: train_batch_50 path: data/train_batch_50-* - split: train_batch_51 path: data/train_batch_51-* - split: train_batch_52 path: data/train_batch_52-* - split: train_batch_53 path: data/train_batch_53-* - split: train_batch_54 path: data/train_batch_54-* - split: train_batch_55 path: data/train_batch_55-* - split: train_batch_56 path: data/train_batch_56-* - split: train_batch_57 path: data/train_batch_57-* - split: train_batch_58 path: data/train_batch_58-* - split: train_batch_59 path: data/train_batch_59-* - split: train_batch_6 path: data/train_batch_6-* - split: train_batch_60 path: data/train_batch_60-* - split: train_batch_61 path: data/train_batch_61-* - split: train_batch_62 path: data/train_batch_62-* - split: train_batch_63 path: data/train_batch_63-* - split: train_batch_64 path: data/train_batch_64-* - split: train_batch_65 path: data/train_batch_65-* - split: train_batch_66 path: data/train_batch_66-* - split: train_batch_67 path: data/train_batch_67-* - split: train_batch_68 path: data/train_batch_68-* - split: train_batch_69 path: data/train_batch_69-* - split: train_batch_7 path: data/train_batch_7-* - split: train_batch_70 path: data/train_batch_70-* - split: train_batch_71 path: data/train_batch_71-* - split: train_batch_72 path: data/train_batch_72-* - split: train_batch_73 path: data/train_batch_73-* - split: train_batch_74 path: data/train_batch_74-* - split: train_batch_75 path: data/train_batch_75-* - split: train_batch_76 path: data/train_batch_76-* - split: train_batch_8 path: data/train_batch_8-* - split: train_batch_9 path: data/train_batch_9-* - split: test_batch_1 path: data/test_batch_1-* - split: test_batch_10 path: data/test_batch_10-* - split: test_batch_11 path: data/test_batch_11-* - split: test_batch_12 path: data/test_batch_12-* - split: test_batch_13 path: data/test_batch_13-* - split: test_batch_14 path: data/test_batch_14-* - split: test_batch_15 path: data/test_batch_15-* - split: test_batch_16 path: data/test_batch_16-* - split: test_batch_17 path: data/test_batch_17-* - split: test_batch_18 path: data/test_batch_18-* - split: test_batch_19 path: data/test_batch_19-* - split: test_batch_2 path: data/test_batch_2-* - split: test_batch_3 path: data/test_batch_3-* - split: test_batch_4 path: data/test_batch_4-* - split: test_batch_5 path: data/test_batch_5-* - split: test_batch_6 path: data/test_batch_6-* - split: test_batch_7 path: data/test_batch_7-* - split: test_batch_8 path: data/test_batch_8-* - split: test_batch_9 path: data/test_batch_9-* --- # AstroPT Euclid Dataset A large-scale multi-band astronomical imaging and photometric dataset from the Euclid Q1 dataset, designed for training deep learning models for astronomical analysis. ## Overview This dataset contains **331,759** Euclid bright and large galaxies (HE<22.5 & segmentation_area>0) with simultaneous observations in visible and near-infrared bands from the Euclid mission, complemented by optical photometry from ground-based surveys. It is specifically curated for the ([AstroPT](https://github.com/Smith42/astroPT)) framework for multi-modal learning tasks in astronomy. **Citation**: Euclid Collaboration: Siudek, M et al. 2025 ([arXiv:2503.15312](https://ui.adsabs.harvard.edu/abs/2025arXiv250315312E/abstract)) ## Data Schema ## Images - **`VIS_image`** (224×224): Visible-band imaging from Euclid VIS instrument (0.55–0.90 μm) - **`NISP_Y_image`** (224×224): Near-infrared Y-band from Euclid NISP (1.0–1.2 μm) - **`NISP_J_image`** (224×224): Near-infrared J-band from Euclid NISP (1.2–1.4 μm) - **`NISP_H_image`** (224×224): Near-infrared H-band from Euclid NISP (1.4–2.0 μm) - **`RGB_image`**: RGB composite image ### Photometry - **`SED`**: Spectral Energy Distribution from optical/NIR photometric filters: 1. u-band (CFHT MegaCam) 2. g-band (Subaru HSC) 3. r-band (CFHT MegaCam) 4. i-band (Pan-STARRS) 5. z-band (Subaru HSC) 6. g-band (DECam) 7. r-band (DECam) 8. i-band (DECam) 9. z-band (DECam) 10. VIS-band (Euclid) 11. Y-band (Euclid NISP) 12. J-band (Euclid NISP) 13. H-band (Euclid NISP) ### Metadata - **`object_id`** (int64): Unique astronomical object identifier ## Quick Start ### Load Dataset (Streaming Mode - Recommended) ```python from datasets import load_dataset # Load a single training batch without caching dataset = load_dataset( "msiudek/astroPT_euclid_dataset", split="train_batch_1", streaming=True ) # Iterate through samples for sample in dataset: object_id = sample['object_id'] vis_image = sample['VIS_image'] # 224×224 array nisp_y = sample['NISP_Y_image'] # 224×224 array nisp_j = sample['NISP_J_image'] # 224×224 array nisp_h = sample['NISP_H_image'] # 224×224 array sed = sample['SED'] # 13 photometric fluxes # Your processing logic here ``` ### Load Multiple Batches ```python from datasets import load_dataset from huggingface_hub import list_repo_files # Get all available splits files = list_repo_files("msiudek/astroPT_euclid_dataset", repo_type="dataset") parquets = [f for f in files if '.parquet' in f] batches = set(f.split('/')[-1].rsplit('-', 2)[0] for f in parquets) # Process all training batches train_batches = sorted([b for b in batches if 'train_batch' in b]) for batch_name in train_batches: try: dataset = load_dataset( "msiudek/astroPT_euclid_dataset", split=batch_name, streaming=True ) for sample in dataset: # Process sample pass except Exception as e: print(f"Error loading {batch_name}: {e}") ``` ### Load Non-Streaming (Download to Disk) ```python dataset = load_dataset( "msiudek/astroPT_euclid_dataset", streaming=False, cache_dir="/path/to/cache" ) ``` ## Data Access & Citation **Framework**: [AstroPT GitHub Repository](https://github.com/Smith42/astroPT) - Multi-modal learning framework - Pre-training and fine-tuning code - Example notebooks and tutorials **Metadata**: [astroPT_euclid_metadata](https://huggingface.co/datasets/msiudek/astroPT_euclid_metadata) - Morphological properties from Euclid - Physical properties (stellar mass, SFR, etc.) - Spectroscopic redshifts from DESI (~5% of sample) - **Citation**: ```bibtex @article{Siudek2025, title={AstroPT: Astronomical Physics Transformers for Multi-modal Learning}, author={Siudek, M and others}, journal={Euclid Collaboration}, eprint={2503.15312}, archivePrefix={arXiv}, year={2025}, url={https://ui.adsabs.harvard.edu/abs/2025arXiv250315312E/abstract} } ``` **Last Updated**: December 2025 **Dataset Version**: 1.0 **Euclid Release**: Q1 **Data License**: CC-BY-4.0 ## Acknowledgments This dataset was generated using data from the **Euclid mission**, an ESA medium-class astronomy and astrophysics space mission. Euclid is a space mission of the European Space Agency (ESA).