Hello Community,
I am seeking feedback on the documentation and workflow implementation for the TESSERA Earth Observation Foundation Model (Cambridge, 2025).
The project focuses on a reproducible Master-Worker architecture designed for resource-constrained environments (Google Colab/L4 GPUs). By applying this to a West African study site (2020–2025), I conducted an ablation-style experiment on cloud-cover thresholds.
Key Finding: I observed a clear 'information ceiling' where increasing scene counts beyond a specific threshold yielded diminishing returns in improving the results. Specifically, while a strict 20% filter resulted in structural breakdown due to low scene counts, a 35% threshold achieved multi-year convergence across the 128-dimensional embeddings.
I have two questions for the group:
- Technical Review: Does the attached README clearly communicate the trade-offs between compute-heavy inference and the 'Model-as-Data' approach for independent researchers?
- Publication Path: I am considering expanding this into a formal technical note or "Methods" paper. Given the focus on reproducibility and empirical threshold selection, which journals or open-access platforms would be most appropriate for this type of workflow validation?"
Link to the draft text: