Open Access
Issue
A&A
Volume 704, December 2025
Article Number A150
Number of page(s) 18
Section Planets, planetary systems, and small bodies
DOI https://doi.org/10.1051/0004-6361/202555598
Published online 15 December 2025

© The Authors 2025

Licence Creative CommonsOpen Access article, published by EDP Sciences, under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

This article is published in open access under the Subscribe to Open model. Subscribe to A&A to support open access publication.

1 Introduction

The current census of extrasolar planets contains more than 7000 objects1, primarily compiled via transit and/or radial velocity (RV) detection methods. These techniques have helped expand our understanding of key physical properties of exoplanets, such as their radius and mass, enabling studies of their bulk density and interior structure (e.g., Seager et al. 2007). However, both the transit and RV methods suffer from known observational biases that favor the detection of close-in orbital companions (e.g., Fischer et al. 2014). To reach longer period orbits, the astrometric detection method can be used (e.g., Binnendijk 1960; Quirrenbach 2011). This involves detecting the reflex motion (wobble) of a star due to an orbiting companion (e.g., exoplanet). The size of this wobble, also known as astrometric orbit signature, scales inversely with distance; therefore, even nearby systems (d<100 pc) yield very small signatures (sub milli-arcsecond). This has resulted in just a handful of brown dwarf or exoplanet detections using astrometry to date (Benedict et al. 2002; Pravdo et al. 2005; Sahlmann et al. 2013; Curiel et al. 2022). To this end, the ESA Gaia telescope (Gaia Collaboration 2016) has provided a unique opportunity for the astrometric detection and characterization of substellar companions to stars. Thanks to its unprecedented precision (end-of-nominal mission estimates for parallax and proper motion are 5 μ as and 3.5 μ as yr−1, respectively, for G= 3-12 mag de Bruijne et al. 2014), Gaia is expected to detect possibly thousands of new planetary and brown dwarf companions in long period orbits down to at least the giant planetary mass regime (Casertano et al. 2008; Sozzetti et al. 2014; Perryman et al. 2014) and up to a few hundred parsec over the whole sky.

In Gaia Data Release 3 (Gaia Collaboration 2023c, hereafter, DR3), the Gaia non-single-star pipelines (Halbwachs et al. 2023; Holl et al. 2023) have produced ∼170 000 orbital solutions for binary systems, providing a first glimpse of the great potential of Gaia for stellar multiplicity analysis (Gaia Collaboration 2023a). A small subset of these solutions (∼1800) correspond to sources with companion candidates in the substellar mass regime (m<80 MJup) (see Sections 5 and 8.1 in Gaia Collaboration 2023a). However, as these authors remark, Gaia DR3 processing limitations do not allow new detections in the planetary mass regime (m<20 MJup) to be claimed solely on the basis of astrometry. The next major Gaia data release, Gaia DR4, is expected to reach at least an order of magnitude higher astrometric precision than Gaia DR3, enabling the expansion of the sensitivity down to the planetary mass regime. Until then, Gaia DR3 can provide important insights onto the demographics of the substellar mass range covering 20 MJupm ≲ 80 MJup (see e.g., Stevenson et al. 2023; Unger et al. 2023). The rich dataset available in DR3, is also well suited for machine learning applications (Guiglion et al. 2024; van Groeningen et al. 2023; Ranaivomanana et al. 2025) and, since the astrometric time series measurements used to produce the DR3 astrometric solutions are not yet publicly available, machine learning can complement the classical orbital fitting method for identification of substellar companions (e.g., Sahlmann & Gómez 2025).

In this work, we present a deep learning framework for identifying companion candidates in Gaia DR3 astrometry. Although deep learning has been extensively applied to transit detection (Shallue & Vanderburg 2018; Ansdell et al. 2018; Dattilo et al. 2019; Valizadegan et al. 2022; Dvash et al. 2022), its use in astrometric companion detection remains largely unexplored. Our method is designed to (i) select those DR3 parameters that effectively trace binarity; (ii) model their complex non-linear relationships using a deep neural network, while mitigating any over-fitting; and (iii) ensure model interpretability, so that predictions can be linked to astrophysical indicators. The resulting model, called ExoDNN, identifies non-single star candidates in Gaia DR3 with high reliability using astrometric fit-quality statistics and other DR3 parameters available in the main source table (Hambly et al. 2022).

The paper is structured as follows. Section 2 introduces the ExoDNN algorithm and describes the data and methods used for its training and validation. In Section 3, we describe how we applied ExoDNN to a subset of Gaia DR3 stars and present a preliminary list of candidate stars to host companions. Section 4 discusses the results, while Section 5 summarizes the conclusions, highlighting synergies with upcoming exoplanet missions and outlining prospects for follow-up observations.

2 The ExoDNN algorithm

2.1 Detection probability and statistical indicators

The astrometric signature of a binary system is typically defined as an angular measure α(arcsecond)=a0/d, where a0 is the semimajor axis of the orbit described by the star around the barycentre of the system (in AU) and d is the distance to the observer (in pc). To determine the probability p of detecting this signature, we can establish a detection criterion based on achieving a specific signal-to-noise ratio (S/N=α/σm), where σm is the measurement uncertainty (ignoring other possible sources of noise). The S/N will then effectively be a cut-off in probability space, that is, serve as a criterion for acceptable confidence level in the detection (and, thus, pS/N). For the case of unknown orbital periods, Casertano et al. (2008) used this criterion and derived a threshold of S/N > 3, for a successful astrometric orbit detection at 95% confidence level. Using a similar criterion, Sahlmann et al. (2015) took into account the number of measurements (Nm) to show that a Nm×S/N20$\sqrt{N_{m}} \times S / N \geq 20$ is necessary for detection when using astrometric data only.

However, the astrometric signature (α) is not directly observable by Gaia. Instead, Gaia measures the one-dimensional (1D) position (abscissa) of a star in the along-scan direction, which we hereafter refer to as η. Gaia data reduction pipeline, the Astrometric Global Iterative Solution (Lindegren et al. 2012, 2021, AGIS, hereafter), uses a single star model to compute the expected along-scan position of a source in the sky at time t (ηexp) and then minimizes the astrometric residuals (observed minus expected) (ηobsηexp) to generate an astrometric solution. This astrometric solution provides the estimated (best-fit) astrometric parameters of the source, along with a set of fit quality statistics (see Table 1). In a binary system, the excess motion induced by the unseen companion will affect the observed position of the star ηobs, causing the astrometric residuals to be larger than in the single star case and resulting in increased values of fit χ2. This has been used by several authors in the past to identify solutions that deviate from the single star model used by AGIS. As an example Belokurov et al. (2020) and Penoyre et al. (2022) used the re-normalized unit weight error (ruwe), which is proportional to the χ2 of the astrometric fit, to search for astrometric binaries in the Gaia catalogue. Stassun & Torres (2021) found a strong correlation between this statistic and the photocentric orbit size of eclipsing binaries. The astrometric excess noise (ε) has also been used by Gandhi et al. (2022) in search for X-ray binaries.

Table 1

AGIS astrometric fit quality statistics.

However, it should be noted that variability (Belokurov et al. 2020), stellar crowding (Luna et al. 2023), or the presence of a proto-planetary stellar disk (Fitton et al. 2022) has been shown to contribute to high values of ruwe without a binary cause. Besides, both statistics are known to be affected by calibration errors that make them sensitive to the magnitude, color of the source (etc.); although these effects have been mitigated by construction in ruwe, they still affect ε in DR3 (Lindegren et al. 2021, Section 5.3). Therefore, these caveats must be taken into account when using AGIS statistics as probes for the presence of unseen companions.

2.2 Deep learning

Despite the limitations mentioned earlier, it is clear that binarity has an effect on the fit quality statistics provided by AGIS, so these can be used to predict the probability, p, that a source from Gaia DR3 is non-single, given an array () of values for those statistics. Now, since we lack an explicit form of the relation between the astrometric signature and the fit quality statistics, finding an analytical form of the probability p(), can be challenging. Therefore, we decided to take a machine learning approach and use a special type of artificial neural network, a deep neural network (DNN), to model p().

In general, artificial neural networks can approximate any continuous function to an arbitrary degree of accuracy (Hornik et al. 1989), and learn such an approximation from data, using provided examples in a setting called supervised learning. If we use a dataset, 𝒟={(1, y1), …,(k, yk)}, composed by pairs of examples (i, yi), where i is array of AGIS fit statistics with known values for a source and yi only has two possible values (yi=0 representing a single star, and yi=1 representing a system with a companion detectable by Gaia), we can teach the network the underlying mapping between and y. The probability density of the random variable y can be described by a continuous Bernoulli distribution (Loaiza-Ganem & Cunningham 2019), expressed as p(yx;θ)=θx(1-θ)1-x,$p(y \mid x ; \theta)=\theta^{x}(1-\theta)^{1-x},$(1) where θ ∈[0, 1] is an unknown parameter that maximizes the probability that y=yi given i, and i ∈[0, 1]N. The DNN defined in this manner will effectively learn the parameter of this distribution and produce the output we are looking for, namely, an estimate of the probability p() over many Bernoulli trials, each with a probability expressed as θ.

thumbnail Fig. 1

Left: simulated orbital periods (black) compared to the orbital periods from DR3 NSS Orbital solutions (shade), with the DR3 time baseline marked in vertical. Right: simulated secondary masses (black), compared to the catalogue of DR3 binary masses (shade), with the peak of the distribution of simulated secondary masses in vertical.

2.3 Generation of the training set

To train ExoDNN, we need example astrometric solutions for the positive case (stars with a detectable companion) and for the negative case (stars with no detectable companion). Given that a homogeneous, sizeable sample of bona fide single stars was not available for our study, we resorted to compiling a set of synthetic data.

We generated 100 000 examples with roughly equivalent proportion (∼50%) of single stars and detectable binary systems, to avoid biasing the neural network towards preferential detection of stars from an over-represented group. These systems were placed randomly in the sky at uniform distances in the range (1 ≤ d ≤ 100 pc). Primary and secondary masses were extracted from the log-normally distributed ranges (0.5 MM1 ≤ 1.5 M) and (10 MJupM2 ≤ 150 MJup), respectively; the latter roughly covering a mass range from giant planets to low-mass stars (Kroupa 2001; Spiegel et al. 2011; Chabrier et al. 2014). We note that our secondary masses (see Figure 1) are generally lower than the ∼1800 candidates from the DR3 NSS catalogue of binary masses (Gaia Collaboration 2023a, Section 5). This is intentional, as our goal is to simulate astrometric signatures induced by very low-mass companions. The orbital periods were drawn from the uniform range (0.1 yr ≤ P ≤ 10 yr), which is the maximum baseline of Gaia observations for the full mission. Once total mass and period of the system were defined, these determined the orbit size (a) through Kepler’s third law. The remaining orbital parameters for the Keplerian companion were the eccentricity (e), orbital inclination (i), argument of the periastron (ω), longitude of the ascending node (Ω), and time at the periastron (tp). These parameters were randomly drawn from the uniform distributions: 0 ≤ e ≤ 1, 0 ≤ i ≤ 90°, 0 ≤ ω ≤ 360°, 0 ≤ Ω ≤ 180°, and 0 ≤ tpP.

We then used the Astromet library from Penoyre et al. (2022) to generate astrometric time series for both single and binary simulated system. Similarly to AGIS, Astromet uses a five-parameter (5 p) astrometric model to describe the motion of a single source. This model depends on right ascension (α*= α cos δ), declination (δ), the proper motions (μα*, μδ), and the parallax (ϖ) of the star, expressed as η(t)=[αα0+μα(tt0)]sinϕt+[δδ0+μδ(tt0)]cosϕt+ϖfϖ$\begin{align*}\eta(t)= & {\left[\alpha^{*}-\alpha_{0}^{*}+\mu_{\alpha^{*}}\left(t-t_{0}\right)\right] \sin \phi_{t} }\\ & +\left[\delta-\delta_{0}+\mu_{\delta}\left(t-t_{0}\right)\right] \cos \phi_{t} \\ & +\varpi f_{\varpi}\end{align*}$(2) where (α0*,δ0$\alpha_{0}^{*}, \delta_{0}$) correspond to the on-sky position at time t0, φt is the position angle of the scan at time t, and fϖ is the source along-scan parallax factor (∂η/∂ϖ). φt and fϖ, were obtained from the Gaia scanning law2.

With the along-scan positions defined, we were able to simulate the 1D observations (ηjobs$\eta_{j}^{\text {obs }}$) that Gaia would perform on source j over a time range equivalent to the Gaia DR3 time baseline of 34 months. Each simulated observation can be expressed as ηjobs(t)=ηj(t)+σm$\eta_{j}^{\text {obs }}(t)=\eta_{j}(t)+\sigma_{m}$, where ηj(t) is given by Eq. (2), and σm is the along-scan single measurement uncertainty. The real Gaia observation uncertainty displays certain dependencies on time, the across-scan rate, and magnitude (see Lindegren et al. 2021 Appendix A); however, for simplicity, we chose a Gaussian measurement error with zero mean and 50 μas standard deviation, the latter corresponding to the median formal DR3 measurement uncertainty at G=12 (again Lindegren et al. 2021 Appendix A). We then fit the single star model from Eq. (2) to our simulated observations and generated a five-parameter astrometric solution for each star in our simulation (see example in Figure 2). The least-squares fit implemented by Astromet (see Penoyre et al. 2020 Appendix D), closely follows AGIS implementation, resulting in astrometric solutions that contain the exact same set of fit quality statistics as provided by AGIS. The choice of fitter is therefore very convenient, because a model trained on those astrometric solutions will be applicable to the five-parameter astrometric solutions available in DR3 main source table.

At this stage, however, we only had (5p) best fit astrometric solutions (α^,δ^,μ^α,μ^δ,ϖ^$\hat{\alpha}, \hat{\delta}, \hat{\mu}_{\alpha}, \hat{\mu}_{\delta}, \hat{\varpi}$), along with their uncertainties and the corresponding fit quality statistics for each solution. To avoid biasing the model towards any positional, motion or distance preference, we dropped the five astrometric solution parameters, and only kept their uncertainties and the fit quality statistics. Also, to have more representative data of the actual DR3 parameter space, we decided to add 16 photometric parameters and 4 spectroscopic DR3 parameters, by conditionally sampling distributions of those parameters derived from real DR3 sources. Details and caveats on the derivation of photometric and spectroscopic information are provided in Appendix A and omitted here for brevity.

This process resulted in a labelled dataset composed of 100 000 examples (i, yi), where i was an array of 31 DR3 parameters, and yi was a value (label) denoting either a binary system with a Gaia detectable companion (yi=1) or a single star (yi=0). The final list of parameters used to train our model is provided in Table 2.

thumbnail Fig. 2

Top: Barycentre motion of a simulated single star corresponding to Eq. (2) with best fit astrometric parameters. The observed position of the source is marked with solid white circles, and the 1D along-scan observations with gray circles. Bottom: same, but for a simulated binary system, where the observed position of the source is perturbed by a companion (exaggerated for illustration purposes).

2.4 Preprocessing

We then scaled robustly the training dataset using the Scikit-Learn package (Pedregosa et al. 2011). To do this, we centred each model parameter by subtracting its median and then dividing by its interquartile range. Robust scaling is a commonly used scaling technique (e.g., Géron 2017) which takes into account the possible presence of outliers and avoids that the different scales naturally present in the data bias the model towards a particular parameter. The training dataset was then split using a 70−15−15% scheme. 70% of the whole data containing positive and negative examples were used to train our network (training set), 15% (validation set) were used to tune the network hyperparameters and the remaining 15% (the test or hold-out set) were left aside to check how well the model generalizes, this is, how well it predicts over unseen data. Shuffling of the whole data was also performed before the train-val-split to ensure adequate mixture of positive and negative examples across the train, test and validation data splits.

2.5 Model training

2.5.1 Model architecture

ExoDNN is composed of an input layer followed by three densely connected hidden layers, one dropout layer and the final output layer. The input layer contains 31 neurons to map our 31 DR3 parameters. The hidden layers have 64 neurons each. The neurons on these layers use a rectified linear unit (ReLU) gate, as it speeds the training in relative comparison to other non-linear gates. The role of the dropout layer is to reduce over-fitting (Srivastava et al. 2014), a situation in which a model produces a perfect prediction over the data it was trained against, but fails to generate good predictions when exposed to unseen new data. Finally, to ensure that a valid probability is produced by the network, the output layer contains a single neuron with a sigmoid activation function which always produces a predicted probability value in the range [0, 1]. We explored network architectures with two to five hidden layers. Configurations with two layers did not achieve the required precision, whereas five layers increased training time without measurable performance gains. The number of neurons per layer was varied between 32 and 256, guided by standard heuristics that balance complexity and computational efficiency. A configuration with 64 neurons per layer yielded the most stable convergence and was therefore adopted. A detailed description of many of the concepts presented in this section such as architecture design, network training, activation gates, back propagation, over-fitting and performance metrics, can be found in Goodfellow et al. (2016).

Table 2

Gaia DR3 parameters used to train the model.

Table 3 summarizes the final architecture of ExoDNN, along with the training hyper-parameters (optimizer, learning rate, mini-batch size, dropout rate and stopping criterion). To implement ExoDNN we used the Tensorflow package (TensorFlow 2015).

Table 3

Neural network hyper-parameters.

2.5.2 Model optimization

A neural network learns by iteratively adjusting its internal parameters (weights and biases) through forward and back propagation of the training data. During forward propagation, the network makes predictions (using small batches of data) of the probability that a training example i in the batch belongs to the positive class (yi=1). Then, these predictions are compared to the true class using a loss function that quantifies how far off the predictions are from the actual labels yi. Our loss function is the average negative log-likelihood of the training set, expressed as l(θ)=-1Niyi×log(θ)+(1-yi)×log(1-θ),$l(\theta)=-\frac{1}{N} \sum_{i} y_{i} \times \log (\theta)+\left(1-y_{i}\right) \times \log (1-\theta),$(3) where N is the number of training samples and yi is the actual class of the i-th sample. The variable θ is the probability computed by the network that the training sample i belongs to the positive class. We note that minimizing the negative loglikelihood is equivalent to maximizing the likelihood of the training data under our parametrized model of the random variable y. During back-propagation, the weights are re-adjusted to lower the loss on the training set. The algorithm responsible for re-adjusting the weights is called optimizer and we chose the Adam optimizer (Kingma & Ba 2015) due to it’s computational efficiency and fast convergence. The model performs multiple forward and backward passes on iterations called epochs. During an epoch a full pass through the entire training set is performed, and this process is repeated until a stopping criterion is met. In our case, we used a stopping criterion called “early stopping” that halts the training when it detects no improvement on the loss computed over the validation set after a successive number of epochs. This choice was also aimed at avoiding a network over-fitting.

The training process explained earlier is designed to teach ExoDNN to produce probability estimates that are as close as possible to the true underlying class probabilities. Therefore, to evaluate the model performance, we selected a metric called the Brier-score, which is equivalent to the mean-squared error but for a probabilistic classification. This metric has values higher or equal than zero, and is defined as Brier=1Ni=1N(p^iyi)2,$Brier=\frac{1}{N} \sum_{i=1}^{N}\left(\hat{p}_{i}-y_{i}\right)^{2},$(4) where N is the number of examples, i is the probability computed by our model for example, i, and yi is the actual value of the random variable y for that example. Figure 3 shows the evolution of the log-loss (left) and brier-score (right) during the training. Since ExoDNN outputs probabilities, to evaluate its performance over a test set with “hard” labels (positive or negative class label), we first needed to establish a classification threshold. This threshold determines when a given predicted probability i belongs to the positive class (binary star). The optimal threshold was computed by making predictions over our hold-out (test) set while varying the probability threshold and finding which of those thresholds maximized a chosen binary classification metric. The selected metric was the f 1-score metric because it represents a trade-off between precision and recall. In this context it is important that we correctly classify stars that host detectable companions as such (precision), and that we capture as many of those as possible (recall). The f 1-score has values between zero and one, and is defined as f1=2*Precision*RecallPrecision+Recall$f 1=\frac{2 * \text { Precision } * \text { Recall }}{\text { Precision }+ \text { Recall }}$(5) ExoDNN retained a very low false positive rate as the classification threshold was varied, as shown in Figure 4 (left), and also maintained high precision and recall at the same time, as shown in Figure 4 (right). The optimal balance between the true positive and the false positive rates over the test data was found for a probability threshold value of p0=0.242. This means that a given example is considered to belong to the positive class, this is, star possibly hosts Gaia detectable companion(s), when the predicted probability on output, , is strictly higher than p0, p̂(i) > p0.

Table 4

Model evaluation results.

thumbnail Fig. 3

Left: evolution of the training and validation set losses over the different epochs. Right: evolution of the Brier-score during the training.

thumbnail Fig. 4

Left: classification threshold computation using the area under curve. Right: classification threshold computation using the precision-recall curve. The optimal threshold that achieves the best compromise between false positive and true positive rates is marked as a red dot. This is our chosen p0=0.242.

2.5.3 Model evaluation

Using the 15% hold-out set from the train-validation-test split, we evaluated the model on unseen data. According to our performance metrics (Section 2.5.2), a well-performing model should exhibit a low Brier score (approaching zero) and a high f 1-score (approaching one). For benchmarking, we compared our model against commonly used binary classifiers: the Logistic Regression (Cox 1958) and the Random Forest (Breiman 2001), a couple of classifiers that are commonly used for binary classification. A good overview of these models is available in Hastie et al. (2001, Chapters 4, 11, 15) and details on their parameter settings are provided in Appendix B. Table 4 shows a summary of the performance metrics obtained by ExoDNN and the alternative classifiers. We found that ExoDNN outperformed the alternative binary classifiers, showing both a low Brier-score and a high f 1-score, supporting its adoption over standard classifiers.

2.6 Model inference and interpretation

We describe above how ExoDNN was built and trained, but we might also consider how it comes up with a predicted probability value for a given input. Our model has learned a non-linear approximation f (using weights and biases) of the true conditional probability, P(y=1 ∣ i), where i is an array of 31 DR3 parameters for a given star. When asked to predict over the input i, it will produce an estimate by computing (i)=σ(f(i)), where σ is the sigmoid function, p^(x¯i)=11+e-f(x¯i).$\hat{p}\left(\bar{x}_{i}\right)=\frac{1}{1+e^{-f\left(\bar{x}_{i}\right)}} .$(6) Now, while the model will use all 31 DR3 parameters to compute , not all of them have the same weight towards that computation. To better understand the relative importance of those 31 DR3 parameters in the prediction, we used SHapley Additive exPlanations (SHAP) (Lundberg & Lee 2017), a commonly used tool to interpret machine learning models. Figure 5 (top) shows the relative importance of the most influential parameters towards the model prediction. The x-axis corresponds to the mean shap value. This is a quantitative measure of how much each feature changes the prediction compared to the average prediction, and gives an idea of its relative importance against other features. We can see how ruwe and astrometric_excess_noise (epsi) have the highest relative importance for ExoDNN prediction, which is expected based on our analysis on how binarity affects the astrometric fit statistics (Section 2.1). Then we have a group of parameters that trace the formal uncertainties in the astrometric solution, the proper motion (declination) and parallax errors (pmDecErr, plxErr), which are also expected to increase due to deviation from a purely linear motion model in the binary case. The RV error (radVelErr) traces the periodic velocity wobble induced by an orbiting companion in the host star, and therefore we would also expect this parameter to have a significant impact on the model decision. Finally, we have a group of photometric uncertainty related parameters (rpFlxOverErr, gFlxOverErr). Sullivan et al. (2025) show how unresolved binarity has a direct impact on flux contamination over the Gaia detection window, resulting in larger photometric residuals. The remaining 23 model parameters were still used, but we note that they have less of an impact in the computation of .

To complement this analysis, we display in Figure 5 (bottom) the computed probability for each of the simulated binaries in the test data sample, along a detection probability proxy computed for different S/N levels. We can see how scales with the astrometric S/N. This is also reassuring with respect to ExoDNN’s behavior, indicating that the model has learned a proxy for the detection probability based on the astrometric S/N ratio noted in Section 2.1.

Table 5

Model testing results.

thumbnail Fig. 5

Top: parameter impact on the model output as estimated by the neural network model ranked by decreasing impact. High parameter values are color-coded in red, while low values are displayed in blue. Bottom: predicted probability for each binary star example on the test dataset (black dots), together with a detection probability proxy computed as p=1−Φ(S/N, δ)), where Φ is the CDF of the normal distribution, S/N=α/σ, and δ is a variable S/N threshold from 1 to 3.

2.7 Model testing

The hold-out dataset used to check ExoDNN generalization (see Section 2.5), is fully synthetic. However, such data cannot capture the intrinsic DR3 calibration features mentioned in Section 2.1. Therefore, we used unresolved binary systems detected by Gaia in DR3 (true positives) to test the behavior of ExoDNN when exposed to unseen and real data. These are: (1) the 384 DR3 NSS OrbitalTargetedSearch* and OrbitalAlternative* solutions (NSS ORBT) from Gaia Collaboration (2023a, Table 1), within 100 pc; (2) the 1332 DR3 NSS Astro-Spectroscopic solutions (combined single lined spectroscopic and astrometric binary) (NSS ASB1) from Gaia Collaboration (2023a) Table 1, within 100 pc; (3) the 48 M-dwarfs from the CARMENES input catalogue (Cifuentes et al. 2025) that have have a corresponding DR3 NSS solution. For all these systems, we had access to an existing orbital solution either from literature or from the Gaia DR3 NSS solutions, with an estimate of the size of the semimajor axis of the photocentric orbit (a0). We then adopted the astrometric S/N detection criterion from Sahlmann et al. (2015), namely: Nobsα/ση>20$\sqrt{N_{\text {obs }}} \alpha / \sigma_{\eta}>20$, where α=a0/d, Nobs corresponds to the astrometric_n_good_observations_al DR3 parameter of the given source, and ση=150 μ as is the median along-scan measurement uncertainty estimated for DR3 sources at G=12 (see Figure 3 in Holl et al. 2023).

This computation is merely a consistency check, since by construction all the selected real sources induce enough astrometric signal to be detectable by Gaia. Therefore, the number of expected detections on the given test sample is the same as the total number of sources for that sample. Finally, to measure ExoDNN baseline false positive rate, we generated a further test sample composed of 5000 new synthetic single stars using the approach described in Section 2.3. A perfect model should predict close to zero probability of hosting a companion for these true negative sources. The false positive rate, as determined from the control sample, resulted in approximately 1.2%.

Table 5 shows the summary results of this process over the entire set of samples. Our model accuracy (detected/expected) varied depending on the sample, from ∼70% for the CARMENES M-dwarf sample, to over 90% for the Gaia NSS single lined astro-spectroscopic binaries.

Figure 6 illustrates how the astrometric signature levels induced by the binary systems in our simulation are (intentionally) comparable to the signatures induced unresolved binaries so far detected by Gaia, and typically higher than currently confirmed exoplanets. Although the overall prediction results are not as optimal as we would like across all real samples, the performance of a model will always decrease when presented with real data. In particular, the characteristic DR3 calibration issues mentioned in Section 2.1 are more likely to affect the fainter and redder M-dwarfs from CARMENES sample. We therefore settled for a version of the model that showed a good compromise between the true and false positive rate, and decided to apply specific post-processing to tackle (at least partially) the mentioned calibration issues. We detail this approach in later sections.

thumbnail Fig. 6

Astrometric signature vs. orbital period for the simulated binary sources (orange), the 384 DR3 NSS ORBT solutions (magenta), the 1332 DR3 NSS ASB1 solutions (blue) and the currently confirmed exoplanets (Schneider et al. 2011) (gray). The red dashed vertical line represents the DR3 time baseline of 34 months in comparison to the shorter periods of the Gaia DR3 unresolved binary orbital solutions. For the DR3 NSS solutions we used the expression α=a0/d to compute the astrometric signature. a0 was computed using the Thiele-Innes elements (A, B, F, G) of the corresponding Gaia DR3 NSS solution and the recipe available in Halbwachs et al. (2023, Appendix A), and d was computed as the inverse of the source parallax. For the simulated sources we used α=M2/M1 × a1/d from Perryman et al. (2014), where M1 and M2 are the primary and secondary masses, respectively, and a1 is the semimajor axis of the orbit of the secondary.

3 Application to Gaia DR3

3.1 ExoDNN application

With our trained and tested deep learning model, we turned to making predictions over real stars from Gaia DR3. We selected a volume limited sample of astrometrically well observed main sequence and giant stars at moderate distances (d<100 pc), from the Gaia DR3 Astrophysical Parameters sample (Apsis, Creevey et al. 2023). This dataset provides estimates for key stellar parameters such as mass, effective temperature and metallicity (among others), which were necessary to characterize the host stars of our potential candidates. To this sample, we applied a set of filtering criteria similar to those used for obtaining the FGKM golden sample of astrophysical parameters (Gaia Collaboration 2023b, CR23 hereafter), but relaxed to reduce the heavy filtering of M stars that takes place in the golden sample. As such, our input for model prediction, fully contains the DR3 FGKM golden sample as a subset, but has a higher fraction of M stars in comparison (see Figure 7). Full details on this process are provided in Appendix C and omitted here for brevity.

This selection resulted in a total of 102117 stars with estimated spectral types F, G, K, and M within 100 pc to which we applied the same pre-processing steps used for the training data, namely the same robust scaling of the model attributes that was described in Section 2.5. The mentioned spectral types correspond to the spectraltype_esphs parameter of the DR3 astrophysical parameters, which is computed using the low-resolution spectra from the BP-RP instrument onboard Gaia. We then used ExoDNN to compute the probability of each of these stars to host unseen companions. Our model returned 14606 stars with a predicted probability p > p0(=0.242), which where considered initial candidates. The breakdown per spectral type (spectraltype_esphs) of these initial candidates is: 6017 M-type stars, 4180 K-type star, 2311 G-type stars and 2098 F-type stars.

thumbnail Fig. 7

Gaia B P minus R P color vs. absolute magnitude for the model input stars (green region). The FGKM golden sample is shown in yellow. Absolute magnitude is computed as: mG = phot_g_mean_mag + 5.0 · log(parallax)−10. Close binary systems where the smaller companion is a brown dwarf or a giant planet are displayed as diamonds (blue) and NSS Orbital* solutions are displayed as squares (magenta). CARMENES sources are displayed as circles (orange). Sources in correspond to the Gaia Catalog of Nearby Stars within d<100 pc.

3.2 Post-processing

We then applied a series of post-processing steps over ExoDNN prediction. These were aimed at obtaining a list of consolidated candidates sources to host unresolved companions. First, we identified candidates that are already known binaries or with confirmed companions by doing a cross-match with a cone search radius of 1.8 arc-seconds against four different catalogues, the 9th Catalog of Spectroscopic Binaries (Pourbaix et al. 2004), the subset of Gaia DR3 NSS orbital solutions (Gaia Collaboration 2023a) within 100 pc, the Washington Double Stars catalogue (Mason et al. 2001) and the online Encyclopedia of Exoplanetary Systems3. This cross-match identified 4383 stars which are present in either one or several of these catalogues. Some illustrative examples of known binaries that ExoDNN model has correctly identified among the initial list of candidates, are 61 Cyg (WDS J21069+3845AB), one of the closest binary systems, Struve 2398 (WDS J18428+5938AB), composed of two M-dwarfs, η Cas (WDS J00491+5749AB), or the spectroscopic binaries 70 Oph (SBC9 1022) and HD137763 (SBC9 1638). Since the model is trained on binary systems without any internal hierarchical structure, it cannot distinguish between binaries and triples or quadruples. Therefore, systems such as κ Tuc (WDS J01158-6853) or σ CrB (WDS 16147+3352), were Gaia has resolved the components, are also reported as candidates by ExoDNN. Other prominent examples of stars within the prediction range of 100 pc with confirmed substellar companions that the model correctly reports as candidates are: ε Eri, a young K-dwarf at ∼3 pc with a confirmed ∼1.6 MJup companion (Benedict et al. 2006), TZAri, a close M5.0V star located at ∼4.8 pc that hosts an RV detected Saturn-type exoplanet (∼0.21 MJup) (Quirrenbach et al. 2022), v And at ∼13.5 pc orbited by several exoplanets including a Jupiter-like companion detected through RV (Curiel et al. 2011), and BD+05 5218 (HIP 117179) a G3V-dwarf 88 pc away that hosts an astrometrically detected ∼44 MJup brown dwarf (Stevenson et al. 2023).

Second, we tackled the calibration issues affecting AGIS statistics (Section 2.1), by following Lindegren et al. (2021), Section 5.3 recommendations. We removed sources with astrometric_excess_noise_sig < 2, and with colors (GB PGR P ≤ 0.4), or (GB PGR P ≥ 4.0). We extended the red end color limit to avoid filtering out known systems which our model correctly considers candidates and would otherwise be filtered out using the recommended constrain of (GB PGR P>3.0). We also removed sources with (phot_g_mean_mag < 6) and (phot_g_mean_mag > 16), as ruwe can be unreliable beyond those magnitude. This filter removed 1579 sources.

Third and final, we used the ipd_frac_multi_peak parameter as an additional discriminator. This statistic is sensitive to the presence of multiple point-spread-functions (PSF) within the same detection window. This situation can occur depending on the scan direction when a partially resolved binary is observed by Gaia. We adopted the suggested value for filtering on this statistic from Fabricius et al. (2021), and only retained sources with ipd_frac_multi_peak<2. Because this parameter is stored as an integer percent, a value of 2 actually means that 0−1% of the transits detected for a given source were multi-peaked. Hence, this filter aims to obtain clean single-peaked sources were the chances of any astrometric perturbation due to a close (partially unresolved) companion of stellar nature is close to zero. This filter removed 1230 sources.

In Figure 8 we illustrate (from left to right) the different cuts in color, magnitude and fraction of multiple PSF found in the detection window, applied on each successive post-processing filter to obtain the final candidates. All together, these three post-processing steps removed 7192 stars from the original candidate list, resulting in a total of 7414 stars which were considered final candidates. The breakdown per spectral type of the final 7414 candidates is: 3757 M-type stars, 1852 K-type star, 950 G-type stars, and 855 F-type stars.

3.3 List of candidates

We show in Figure 9 a color-magnitude diagram of the Gaia Catalog of Nearby Stars (Gaia Collaboration 2021), where we highlight the final 7414 consolidated candidates reported by ExoDNN, along with a few sources to help showcase ExoDNN predictions. These showcase examples are Gaia-4b and Gaia-5b (Stefánsson et al. 2025), a massive planet and a brown dwarf, respectively, with masses of m=11.8 ± 0.7 MJup and m=20.9 ± 0.5 MJup, the only substellar companions from DR3 that have been externally confirmed so far via RV follow-up observations. Both Gaia-4b and Gaia-5b are reported as candidates by ExoDNN. G 15-6, where Sahlmann & Gómez (2025) have recently identified a brown dwarf companion with an astrometric mass m=62.4 ± 6.2 MJup, is also reported as a candidate. We also show Gaia black hole 3 (Gaia Collaboration 2024), whose host star is also detected as a candidate. We omit Gaia-1b and Gaia-2b since these are close-in transiting planets (Panahi et al. 2022) with very short orbital periods (∼3 days) that do not produce enough gravitational pull to be astrometrically detectable in DR3 (or even DR4/5). Another interesting example is HD 164604, a K-dwarf at ∼40 pc orbited by a ∼14 MJup super Jupiter and detected via astrometry (Gaia Collaboration 2023a), that is also reported as a candidate by ExoDNN. We searched for signs of preferential detection of candidates on particular sky positions, distances, Gaia colors, or Gaia magnitudes. Figure 10 shows the distributions of some of those parameters, which do not reveal any significant bias in the model output candidates. When we analyse the nearby candidates reported by ExoDNN, shown in Table 6, we can see that the majority are M-dwarfs. This is not totally surprising since the astrometric signature of an hypothetical companion will be stronger in an M-dwarf due to the higher mass ratio. This, together with their abundant numbers (∼75% of the stellar population Henry & Jao 2024), makes M stars usual targets for RV exoplanet surveys (e.g., Ribas et al. 2023). We find among these closest candidates an illustrative example (HD 52698) on how the model decision is driven by the most influential parameters ruwe and astrometric_excess_noise. This spectroscopic binary has intrinsic high values of both astrometric fit statistics, which drives ExoDNN to make a strong prediction of the probability that this source has companions, 0.99 probability against the average of the other 12 candidates 0.37.

To check whether any biases had been introduced by adopting a deep learning approach in this domain, we compared ExoDNN predictions with those from Sahlmann & Gómez (2025). In that work, the authors used a semisupervised anomaly detection framework that combines feature importance and ensemble classifiers to identify potential companions to stars. The scientific goal (astrometric detection of sub-stellar companions) is equivalent to ours, but the technique is different, thus making the comparison relevant. We fed ExoDNN with the 22 high-probability candidate sources reported by these authors (see their Table 4). To do this, we preprocessed the required model parameters corresponding to each of those 22 Gaia DR3 sources and then computed the probability of each of them to host a substellar companion using ExoDNN. We present the predictions made by ExoDNN over those 22 candidates in Table 7. ExoDNN predicts that 16 out of 22(∼72%) also are candidates to host companions. However, as noted by Sahlmann & Gómez (2025) (see their section 4.2), some of these candidates (e.g., * 54 Cas and BD+75 510), were identified after DR3 release as false positives. Finally, the six sources not considered as candidates by ExoDNN have, in relative comparison to the others, a combination of ruwe and astrometric_excess_noise that results in a lower probability than our established threshold. The rest of the candidates (16) are detected by ExoDNN with computed probabilities to hold companions in the range (0.25 to 0.99) depending on the relative values of the two main astrometric fit quality indicators. We find cases such as HD 207740, HD 40503, GSC 09436-01089 and G15-6, where the relative high values of both the ruwe and astrometric_excess_noise drive ExoDNN prediction to a high predicted probability of holding companions. This result is reassuring, as it represents an independent check of the ExoDNN behavior.

thumbnail Fig. 8

Left: astrometric_excess_noise vs. Gaia color bp_rp. Model inputs are displayed in black and candidates in green. The vertical red lines show the applied color cuts. Orange line corresponds to the running median of the excess noise and the shaded orange region corresponds to the 16th-84th percentiles. Middle: equivalent plot but for ruwe vs. Gaia G band magnitude. The vertical red lines show the applied bright and faint Gaia magnitude cuts. Right: equivalent plot but for ipd_frac_multi_peak vs. Gaia G magnitude. The horizontal red line corresponds to the selected cut-off limit for this parameter.

thumbnail Fig. 9

Gaia BP minus RP color vs. absolute magnitude diagram of Gaia DR3 Catalog of Nearby Sources. The 7414 final candidates reported by ExoDNN after filtering are displayed in green. Prediction for a few reference objects, Proxima Centauri, LHS1610, BD+05 5128, HD 164604, Gaia-3b, Gaia-4b and Gaia-5b, etc., is also shown. Sources with known companions detected by ExoDNN are marked by the green boxes, while red-crosses refer to reference objects for which model predicted probability <p0, that is, undetected. ExoDNN closest new candidates are highlighted to the left of the main sequence with blue circles.

thumbnail Fig. 10

Top: distribution of the right ascension values for candidates sources (solid black line) and the original input for the model (gray) Middle: equivalent display but for Gaia BP minus RP color (bp-rp) distribution. Bottom: equivalent display but for the Gaia magnitude (phot_g_mean_mag) in this case.

4 Discussion

4.1 Effects of filtering

The significant reduction (∼50%) from the original 14606 candidates to the final 7414 has its origin in the post-processing filters we applied (Section 3.2). Over 30% of the original candidates were identified during the cross-match (first filter) as previously known binary or multiple systems. These candidates are considered true positives, which ExoDNN has correctly identified and thus, they were removed because the focus of our study is on finding genuinely new candidates. The remaining filtering steps remove 20% of the initial candidates but affect differently across the spectral types. While the astrometric excess noise significance filter (second filter) affects the fainter (M-type) stars more heavily, the multi peak discriminator (third filter) affects more F, G and K-type stars. The net effect of these filters in the total numbers per spectral type is illustrated in Figure 11.

Another prominent feature of our final candidates, is the apparent depletion of faint candidates (MG>12). This is due to the strong filtering induced by the second post-processing step (Section 3.2) for faint magnitudes. The fraction of DR3 sources with significant (>2) excess noise decreases from 97% at 9 ≤ G ≤ 12 to only 20% at G=14 (Lindegren et al. 2021, Table 4), and therefore that filter is particularly efficient in removal of M-type candidates in the magnitude range (12 ≤ G ≤ 15). These sources frequently exhibit excess noise significance values lower than 2, see Figure 11 (s) values. Also, the significant depletion of ∼50% F and G-type initial candidates is caused by our third filtering step. The general increased multiplicity fraction of F and G dwarfs, ∼60% for F-G stars vs. 25−40% for M-dwarfs (Duchêne & Kraus 2013), will result in relatively higher numbers of the ipd_frac_multi_peak parameters for the spectral types F, G and K over M-type sources, see Figure 11(f) values. Therefore the third filtering step removes primarily the brighter and more massive initial candidates. We plan to improve these features in further versions of ExoDNN by applying refined selection criteria for sources around the lower astrometric excess noise significant limit and for those with multiple psfs within the detection window.

Table 6

ExoDNN closest-candidates prediction.

Table 7

ExoDNN prediction on external candidates.

4.2 Host characterization

Thanks to the selection of the model input from the DR3 Astrophysical Parameters dataset, we have access to stellar parameters such as the mass, metallicity, and age of the candidate stars. These parameters have been derived by the astrophysical parameter inference system (Apsis, Fouesneau et al. 2023), using the low-resolution spectra provided by Gaia’s blue and red photometers (BP/RP), and provide valuable information to characterize our candidate stars population. If we analyse the corresponding distributions of mass, metallicity and age for the candidates reported by ExoDNN, we find a peak mass Mcand.=0.7 ± 0.1 M and a peak metallicity of M/Hcand=−0.1 ± 0.01 dex (see Figure 12). This would suggest that the candidates found by ExoDNN are similar, in terms of stellar properties, to the NSS astro-spectroscopic binaries, but not to the NSS eclipsing binaries (see Gaia Collaboration 2023a). The model has been trained on a distribution of periods that overlaps with the one from the NSS astro-spectroscopic binaries (typically hundreds of days), which results in a model that is more sensitive to those orbital configurations rather than to the much shorter orbital periods of the eclipsing binaries (typically a few days).

It should be noted that the parameters derived by the Apsis may suffer from calibration problems in the case of unresolved binaries (Fouesneau et al. 2023, Section 3.5.2). This is due to the lack of sufficiently high-quality synthetic models of BP and RP spectra of unresolved binaries and, therefore, our comparison can only be regarded as qualitative.

thumbnail Fig. 11

Overall number of candidates per spectral type (spectraltype_esphs), pre-filtering (blue) and post-filtering (orange). Blue text labels on top of each group of bars correspond respectively to (s): median values of the astrometric_excess_ noise_sig parameter for the given spectral type and (f): median values of the ipd_frac_multi_peak parameter for the given spectral type.

4.3 Caveats

We remark that in its current version, ExoDNN reports candidate stars to host companions detectable by Gaia, but it cannot assess the nature (stellar or substellar) of the potential companions. In general, when the current ExoDNN reports a source as a candidate, we may face different scenarios. One, where the signatures of existing companions in the system are of the order of the DR3 measurement uncertainty. In this case, ExoDNN prediction may be just detecting the signatures of existing companions in the system or alternatively, indicating the presence of a yet unknown companion in a longer period orbit. Another possible scenario occurs when the source has no reported companions yet, but ExoDNN predicts their presence. And yet another scenario may occur, which represents an astrophysical false positive. In this latter case, the AGIS statistics are inflated but for reasons other than multiplicity, triggering a false positive. Irrespective of the case, supporting evidence such as additional RV measurements, epoch astrometry or imaging data, are necessary to first confirm or rule out the existence of companions, and only then be able to provide estimates on their masses.

thumbnail Fig. 12

Top: primary mass distribution of the candidate stars reported by ExoDNN (green). Also added for comparison are the mass distributions corresponding to the NSS Orbital* (orange), AstroSpectroSB1 (blue), and Eclipsing* binary (black) solutions from the DR3 NSS. The ExoDNN predictions show the 6% model false positive rate as downward arrows. Bottom: equivalent distributions but for metallicity of the candidates reported by ExoDNN.

5 Conclusions

Using the rich dataset provided in Gaia DR3, we created a deep learning model, ExoDNN, that uses the available DR3 parameters to predict the probability of a Gaia DR3 star to host one or more companions. To do this, ExoDNN takes advantage of the existing correlation between AGIS fit quality statistics and the deviation from a single source model. We have shown how ExoDNN is able to correctly detect different types of multiplicity, from known binaries on wide orbits to star+brown dwarf configurations. When applied to a sample of stars from the Gaia DR3 catalogue within 100 pc, ExoDNN generated a list of 7414 new candidates to host unresolved companions. The number of candidate stars detected is comparable in order of magnitude to predictions from earlier studies (e.g., Casertano et al. 2008; Perryman et al. 2014). The post-processing steps we applied are conservative, which resulted in the removal of approximately ∼50% of the initial candidates found by ExoDNN, but this step was necessary to avoid contamination from sources possibly affected by DR3 AGIS calibration limitations. A false positive rate of ∼1.2% is expected from the current version of ExoDNN and should be taken into account when assessing the proposed candidates by performing additional scrutiny on a per-source basis. The main limitation of the current version of ExoDNN is the inability to determine the nature (stellar or substellar mass) of potential companions without incorporating external data, such as RV measurements or epoch astrometry. In addition, since the ExoDNN predictions are generated from DR3 astrometry, those predictions are limited by the available precision in DR3 measurements, which has been the main driver in applying ExoDNN to volume-limited sample of stars (d<100 pc). Technically speaking, however, nothing prevents applying the model to sources further away (one extreme example is ExoDNN prediction over Gaia BH3, located at d ∼590 pc). Despite these limitations and until Gaia DR4 is published along with a list of exoplanet candidates, ExoDNN candidates could be a powerful tool for the community for benchmarking different theories of planetary formation given its statistical relevance. In addition, the proposed solar-type (G) candidates could be interesting in the context of future missions such PLATO (Rauer et al. 2022). In Figure 13, we display the G-type candidates found by ExoDNN over the PLATO first field of view, FoV South 1 (240 G-dwarf candidates).

We plan to further enhance the current model performance in several ways. First, we would like to improve the model sensitivity specially over the faint end covering M-dwarfs. Training ExoDNN on additional examples of real M-type stars that are known to host companions could help it learn (at least partially), the intrinsic Gaia DR3 calibration artifacts affecting these sources. This, in turn, could eliminate the current need for applying specific excess-noise cuts to M-type candidates. Then, we will also revise the current model architecture to include uncertainty estimations. This will introduce the ability to inform the user on the uncertainty in a given probability prediction and require the evolution of the current neural network architecture towards a deep probabilistic architecture (e.g., Neal 2012; Blundell et al. 2015; Lakshminarayanan et al. 2017; Abdar et al. 2021). Last but not least, to gain the ability to asses the nature of reported candidates, we will need to add external datasets, such as RV values or astrometric time series. This will allow us to constrain the candidate companion masses and increase the current scope of the model from astrometric detection of substellar companions to a broader one, namely, the characterization of substellar companions.

Some of our candidates are close enough to make them suitable targets for follow-up observations by different means. For candidates at moderate distances, RV follow-up observations with high-resolution spectrographs such as HARPS (Cosentino et al. 2012) or ESPRESSO (Pepe et al. 2021), would allow for the presence of companions to be confirmed or for orbital configuration and masses to be at least partially constrain. We intend to asses the suitability for follow-up observations of the most promising candidates reported in this work and identify the best follow-up strategy.

thumbnail Fig. 13

ExoDNN G-type star candidates overlay in PLATO first field of view (LOPS1).

Data availability

The full version of the 7414 new Gaia DR3 candidate stars to host companions is available at the CDS via https://cdsarc.cds.unistra.fr/viz-bin/cat/J/A+A/704/A150

Acknowledgements

C.C. acknowledges financial support by the Consejo Superior de Investigaciones Científicas (CSIC) through the internal project 2023AT003 associated to the RYC2021-031640-I. J.L.-B. is funded by the Spanish Ministry of Science and Universities (MICIU/AEI/10.13039/501100011033) and NextGenerationEU/PRTR grants PID2019-107061GB-C61, PID2021-125627OB-C31, CNS2023-144309, and PID2023-150468NB-I00. This research made use of the following software: astropy, (a community-developed core Python package for Astronomy, Astropy Collaboration 2022), SciPy (Virtanen et al. 2020), matplotlib (a Python library for publication quality graphics Hunter 2007), numpy (Harris et al. 2020) and Tensorflow (TensorFlow 2015). This research has made use of data from the European Space Agency (ESA) mission Gaia (https://www.cosmos.esa.int/gaia), processed by the Gaia Data Processing and Analysis Consortium (DPAC, https://www.cosmos.esa.int/web/gaia/dpac/consortium). Funding for the DPAC has been provided by national institutions, in particular the institutions participating in the Gaia Multilateral Agreement. We also made use of NASA’s Astrophysics Data System (ADS) Bibliographic Services, the SIMBAD database, operated at CDS, and the NASA Exoplanet Archive, which is operated by the California Institute of Technology, under contract with the National Aeronautics and Space Administration under the Exoplanet Exploration Program. We also used the data obtained from or tools provided by the portal exoplanet.eu of The Extrasolar Planets Encyclopaedia.

References

  1. Abdar, M., Pourpanah, F., Hussain, S., et al. 2021, Information Fusion, 76, 243 [CrossRef] [Google Scholar]
  2. Alonso-Floriano, F. J., Morales, J. C., Caballero, J. A., et al. 2015, A&A, 577, A128 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  3. Ansdell, M., Ioannou, Y., Osborn, H. P., et al. 2018, ApJ, 869, L7 [NASA ADS] [CrossRef] [Google Scholar]
  4. Astropy Collaboration (Price-Whelan, A. M., et al.,) 2022, ApJ, 935, 167 [NASA ADS] [CrossRef] [Google Scholar]
  5. Belokurov, V., Penoyre, Z., Oh, S., et al. 2020, MNRAS, 496, 1922 [Google Scholar]
  6. Benedict, G. F., McArthur, B. E., Forveille, T., et al. 2002, ApJ, 581, L115 [NASA ADS] [CrossRef] [Google Scholar]
  7. Benedict, G. F., McArthur, B. E., Gatewood, G., et al. 2006, AJ, 132, 2206 [Google Scholar]
  8. Binnendijk, L., 1960, Properties of Double Stars: A Survey of Parallaxes and Orbits (Philadelphia: University of Pennsylvania Press) [Google Scholar]
  9. Blundell, C., Cornebise, J., Kavukcuoglu, K., & Wierstra, D., 2015, in International conference on machine learning, PMLR, 1613 [Google Scholar]
  10. Breiman, L., 2001, Mach. Learn., 45, 5 [Google Scholar]
  11. Bressan, A., Marigo, P., Girardi, L., et al. 2012, MNRAS, 427, 127 [NASA ADS] [CrossRef] [Google Scholar]
  12. Casertano, S., Lattanzi, M. G., Sozzetti, A., et al. 2008, A&A, 482, 699 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  13. Chabrier, G., Johansen, A., Janson, M., & Rafikov, R., 2014, in Protostars and Planets VI, eds. H. Beuther, R. S. Klessen, C. P. Dullemond, & T. Henning (Tucson: University of Arizona Press), 619 [Google Scholar]
  14. Cifuentes, C., Caballero, J. A., González-Payo, J., et al. 2025, A&A, 693, A228 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  15. Cosentino, R., Lovis, C., Pepe, F., et al. 2012, SPIE Conf. Ser., 8446, 84461V [Google Scholar]
  16. Cox, D. R., 1958, J. R. Stat. Soc., 20, 215 [Google Scholar]
  17. Creevey, O. L., Sordo, R., Pailler, F., et al. 2023, A&A, 674, A26 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  18. Curiel, S., Cantó, J., Georgiev, L., Chávez, C. E., & Poveda, A., 2011, A&A, 525, A78 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  19. Curiel, S., Ortiz-León, G. N., Mioduszewski, A. J., & Sanchez-Bermudez, J., 2022, AJ, 164, 93 [Google Scholar]
  20. Dattilo, A., Vanderburg, A., & Shallue, C., 2019, AAS Meeting Abstracts, 233, 140.16 [Google Scholar]
  21. de Bruijne, J. H. J., Rygl, K. L. J., & Antoja, T., 2014, EAS Pub. Ser., 67–68, 23 [Google Scholar]
  22. Duchêne, G., & Kraus, A., 2013, ARA&A, 51, 269 [Google Scholar]
  23. Dvash, E., Peleg, Y., Zucker, S., & Giryes, R., 2022, AJ, 163, 237 [Google Scholar]
  24. Eker, Z., Soydugan, F., Soydugan, E., et al. 2015, AJ, 149, 131 [Google Scholar]
  25. Eker, Z., Soydugan, F., Bilir, S., et al. 2020, MNRAS, 496, 3887 [NASA ADS] [CrossRef] [Google Scholar]
  26. Fabricius, C., Luri, X., Arenou, F., et al. 2021, A&A, 649, A5 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  27. Fischer, D. A., Howard, A. W., Laughlin, G. P., et al. 2014, Exoplanet Detection Techniques (Tucson: University of Arizona Press) [Google Scholar]
  28. Fitton, S., Tofflemire, B. M., & Kraus, A. L., 2022, Res. Notes Am. Astron. Soc., 6, 18 [Google Scholar]
  29. Fouesneau, M., Frémat, Y., Andrae, R., et al. 2023, A&A, 674, A28 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  30. Gaia Collaboration (Prusti, T., et al.,) 2016, A&A, 595, A1 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  31. Gaia Collaboration (Smart, R. L., et al.,) 2021, A&A, 649, A6 [EDP Sciences] [Google Scholar]
  32. Gaia Collaboration (Arenou, F., et al.,) 2023a, A&A, 674, A34 [CrossRef] [EDP Sciences] [Google Scholar]
  33. Gaia Collaboration (Creevey, O. L., et al.,) 2023b, A&A, 674, A39 [CrossRef] [EDP Sciences] [Google Scholar]
  34. Gaia Collaboration (Vallenari, A., et al.,) 2023c, A&A, 674, A1 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  35. Gaia Collaboration (Panuzzo, P., et al.,) 2024, A&A, 686, L2 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  36. Gandhi, P., Buckley, D. A. H., Charles, P. A., et al. 2022, MNRAS, 510, 3885 [CrossRef] [Google Scholar]
  37. Géron, A., 2017, Hands-on Machine Learning with Scikit-Learn and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems (Sebastopol, CA: O’Reilly Media) [Google Scholar]
  38. Goodfellow, I., Bengio, Y., & Courville, A., 2016, Deep Learning (Cambridge, MA: MIT Press) [Google Scholar]
  39. Gosset, E., Damerdji, Y., Morel, T., et al. 2025, A&A, 693, A124 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  40. Guiglion, G., Nepal, S., Chiappini, C., et al. 2024, A&A, 682, A9 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  41. Halbwachs, J.-L., Pourbaix, D., Arenou, F., et al. 2023, A&A, 674, A9 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  42. Hambly, N., Andrae, R., De Angeli, F., et al. 2022, Gaia DR3 Documentation, Chapter 20: Data Model Description, Gaia Data Processing and Analysis Consortium (DPAC), European Space Agency [Google Scholar]
  43. Harris, C. R., Millman, K. J., van der Walt, S. J., et al. 2020, Nature, 585, 357 [NASA ADS] [CrossRef] [Google Scholar]
  44. Hastie, T., Tibshirani, R., & Friedman, J., 2001, The Elements of Statistical Learning, Springer Series in Statistics (New York, NY, USA: Springer New York Inc.) [Google Scholar]
  45. Henry, T. J., & Jao, W.-C., 2024, ARA&A, 62, 593 [Google Scholar]
  46. Holl, B., Sozzetti, A., Sahlmann, J., et al. 2023, A&A, 674, A10 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  47. Hornik, K., Stinchcombe, M., & White, H., 1989, Neural Netw., 2, 359 [NASA ADS] [CrossRef] [Google Scholar]
  48. Hunter, J. D., 2007, CSE, 9, 90 [Google Scholar]
  49. Keenan, P. C., & McNeil, R. C., 1989, ApJS, 71, 245 [Google Scholar]
  50. Kingma, D. P., & Ba, J., 2015, in ICLR (Poster), eds. Y. Bengio, & Y. LeCun [Google Scholar]
  51. Kroupa, P., 2001, MNRAS, 322, 231 [NASA ADS] [CrossRef] [Google Scholar]
  52. Lakshminarayanan, B., Pritzel, A., & Blundell, C., 2017, Advances in Neural Information Processing Systems, 30 [Google Scholar]
  53. Lépine, S., Hilton, E. J., Mann, A. W., et al. 2013, AJ, 145, 102 [Google Scholar]
  54. Lindegren, L., Lammers, U., Hobbs, D., et al. 2012, A&A, 538, A78 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  55. Lindegren, L., Hernández, J., Bombrun, A., et al. 2018, A&A, 616, A2 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  56. Lindegren, L., Klioner, S. A., Hernández, J., et al. 2021, A&A, 649, A2 [EDP Sciences] [Google Scholar]
  57. Loaiza-Ganem, G., & Cunningham, J. P., 2019, Advances in Neural Information Processing Systems, 32 [Google Scholar]
  58. Luna, A., Marchetti, T., Rejkuba, M., & Minniti, D., 2023, A&A, 677, A185 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  59. Lundberg, S. M., & Lee, S.-I., 2017, Advances in Neural Information Processing Systems, 30 [Google Scholar]
  60. Mason, B. D., Wycoff, G. L., Hartkopf, W. I., Douglass, G. G., & Worley, C. E., 2001, AJ, 122, 3466 [Google Scholar]
  61. Morton, T. D., 2015, Astrophysics Source Code Library [record ascl:1503.010] [Google Scholar]
  62. Neal, R. M., 2012, Bayesian Learning for Neural Networks (Springer Science & Business Media), 118 [Google Scholar]
  63. Newton, E. R., Charbonneau, D., Irwin, J., et al. 2014, AJ, 147, 20 [Google Scholar]
  64. Panahi, A., Zucker, S., Clementini, G., et al. 2022, A&A, 663, A101 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  65. Pedregosa, F., Varoquaux, G., Gramfort, A., et al. 2011, JMLR, 12, 2825 [Google Scholar]
  66. Penoyre, Z., Belokurov, V., Evans, N. W., Everall, A., & Koposov, S. E., 2020, MNRAS, 495, 321 [Google Scholar]
  67. Penoyre, Z., Belokurov, V., & Evans, N. W., 2022, MNRAS, 513, 5270 [Google Scholar]
  68. Pepe, F., Cristiani, S., Rebolo, R., et al. 2021, A&A, 645, A96 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  69. Perryman, M., Hartman, J., Bakos, G. Á., & Lindegren, L. 2014, ApJ, 797, 14 [Google Scholar]
  70. Pourbaix, D., Tokovinin, A. A., Batten, A. H., et al. 2004, A&A, 424, 727 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  71. Pravdo, S. H., Shaklan, S. B., & Lloyd, J., 2005, ApJ, 630, 528 [Google Scholar]
  72. Quirrenbach, A., 2011, Exoplanets, ed. S Seager (Tucson, AZ: University of Arizona Press), 526, 157 [Google Scholar]
  73. Quirrenbach, A., Passegger, V. M., Trifonov, T., et al. 2022, A&A, 663, A48 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  74. Ranaivomanana, P., Uzundag, M., Johnston, C., et al. 2025, A&A, 693, A268 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  75. Rauer, H., Aerts, C., Deleuil, M., et al. 2022, European Planetary Science Congress, 453 [Google Scholar]
  76. Reid, I. N., Hawley, S. L., & Gizis, J. E., 1995, AJ, 110, 1838 [Google Scholar]
  77. Ribas, I., Reiners, A., Zechmeister, M., et al. 2023, A&A, 670, A139 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  78. Sahlmann, J., & Gómez, P., 2025, MNRAS, 537, 1130 [Google Scholar]
  79. Sahlmann, J., Lazorenko, P. F., Ségransan, D., et al. 2013, A&A, 556, A133 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  80. Sahlmann, J., Triaud, A. H. M. J., & Martin, D. V., 2015, MNRAS, 447, 287 [NASA ADS] [CrossRef] [Google Scholar]
  81. Schneider, J., Dedieu, C., Le Sidaner, P., Savalle, R., & Zolotukhin, I., 2011, A&A, 532, A79 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  82. Schweitzer, A., Passegger, V. M., Cifuentes, C., et al. 2019, A&A, 625, A68 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  83. Scott, D. W., 2015, Multivariate Density Estimation: Theory, Practice, and Visualization (Hoboken: Wiley) [Google Scholar]
  84. Seager, S., Kuchner, M., Hier-Majumder, C. A., & Militzer, B., 2007, ApJ, 669, 1279 [NASA ADS] [CrossRef] [Google Scholar]
  85. Shallue, C. J., & Vanderburg, A., 2018, AJ, 155, 94 [NASA ADS] [CrossRef] [Google Scholar]
  86. Shan, Y., Revilla, D., Skrzypinski, S. L., et al. 2024, A&A, 684, A9 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  87. Shkolnik, E., Liu, M. C., & Reid, I. N., 2009, ApJ, 699, 649 [CrossRef] [Google Scholar]
  88. Silverman, B. W., 1986, Density Estimation for Statistics and Data Analysis, Chapman & Hall/CRC monographs on statistics and applied probability (London: Chapman and Hall) [Google Scholar]
  89. Sozzetti, A., Giacobbe, P., Lattanzi, M. G., et al. 2014, MNRAS, 437, 497 [NASA ADS] [CrossRef] [Google Scholar]
  90. Spiegel, D. S., Burrows, A., & Milsom, J. A., 2011, ApJ, 727, 57 [Google Scholar]
  91. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., & Salakhutdinov, R., 2014, JMLR, 15, 1929 [Google Scholar]
  92. Stassun, K. G., & Torres, G., 2021, ApJ, 907, L33 [NASA ADS] [CrossRef] [Google Scholar]
  93. Stefánsson, G., Mahadevan, S., Winn, J. N., et al. 2025, AJ, 169, 107 [Google Scholar]
  94. Stevenson, A. T., Haswell, C. A., Barnes, J. R., & Barstow, J. K., 2023, MNRAS, 526, 5155 [NASA ADS] [CrossRef] [Google Scholar]
  95. Sullivan, K., Kraus, A. L., Berger, T. A., & Huber, D., 2025, AJ, 169, 29 [Google Scholar]
  96. TensorFlow, G., 2015, Google Research, 10, s15326985ep4001 [Google Scholar]
  97. Torres, C. A. O., Quast, G. R., da Silva, L., et al. 2006, A&A, 460, 695 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  98. Unger, N., Ségransan, D., Barbato, D., et al. 2023, A&A, 680, A16 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  99. Valizadegan, H., Martinho, M. J. S., Wilkens, L. S., et al. 2022, ApJ, 926, 120 [NASA ADS] [CrossRef] [Google Scholar]
  100. van Groeningen, M. G. J., Castro-Ginard, A., Brown, A. G. A., Casamiquela, L., & Jordi, C., 2023, A&A, 675, A68 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
  101. Virtanen, P., Gommers, R., Oliphant, T. E., et al. 2020, Nat. Methods, 17, 261 [Google Scholar]

1

Statistics as of early 2025 from: https://exoplanet.eu/home/

2

Gaia scanning law for the full mission can be obtained at: https://www.cosmos.esa.int/web/gaia/full-mission-scanning-law

3

Accessed early May 2025 at: https://exoplanet.eu/home/

Appendix A Adding photometric and spectroscopic information

To add photometric and spectroscopic information for every simulated source we first built two differentiated samples, one corresponding to the single stars (negative case) and another for the binary stars (positive case). Each sample contained a combination of 50000 main sequence and giant real stars. In the absence of a large enough sample of bona fide single stars to construct our single stars subset, we resorted to the Gaia DR3 FGKM golden sample (Gaia Collaboration 2023b). These are set of about three million sources with very-high-quality photometry, astrometry, and stellar parameters. From the full FGKM sample we selected sources with no appreciable signs of binarity or multiplicity. To do this we required well-observed single-star solutions with a low ruwe (ruwe ≤ 1.2 Lindegren et al. 2018; Fabricius et al. 2021) and at least 11 visibility periods. We also added further filtering on ipd_frac_multi_peak to avoid double stars with a large separation (≳ 1.5 arcsecond. Fabricius et al. 2021) and ipd_gof_harmonic_amplitude to also reject pairs with smaller separations. Finally, to minimize contamination from binaries we excluded all sources listed in the Gaia DR3 Non-Single Star (NSS) catalogue (Halbwachs et al. 2023; Gosset et al. 2025) (astrometric and spectroscopic binaries) using the non_single_star flag. We joined the resulting subset with the astrophysical parameters DR3 table gaiadr3.astrophysical_parameters to obtain the mass estimate for each source. This can be translated into the following query in the Gaia archive:

SELECT TOP 100000 A.*, B.* FROM gaiadr3.gold_sample_fgkm_stars AS B
JOIN gaiadr3.gaia_source AS A USING (source_id)
WHERE A.ruwe <1.2
AND A.visibility_periods_used > 11
AND A.ipd_frac_multi_peak < 2
AND A.ipd_gof_harmonic_amplitude < 0.1
AND A.non_single_star =0

For the binary stars we did the opposite operation, selecting astrometrically well-observed sources from the main Gaia DR3 table with high ruwe value. We removed the cuts on initial parameter determination (IPD) parameters and set the non_single_star flag to value 1, which indicates that the source is identified as an astrometric binary in the DR3 NSS catalogue. This can be translated into the following query in the Gaia archive:

SELECT TOP 100000 A.*, B.* FROM gaiadr3.gaia_source AS A
JOIN gaiadr3.astrophysical_parameters AS B USING
(source_id)
WHERE A.ruwe > 1.4
AND A.visibility_periods_used > 11
AND A.non_single_star = 1

A description of all the parameters used in these queries is available in the online Gaia documentation4.

thumbnail Fig. A.1

TOP: Mass distribution of single (blue) and binary (orange) samples with bins of 0.1 M on the variable mass_flame. MIDDLE: Distribution of color indices for single sources (gray) compared to sampled colors (orange), obtained from a 1D KDE fit corresponding to the low end of the simulation mass range. BOTTOM: Equivalent plot as in the middle but for the binary sources in this case.

For each generated sampled we only retained those stars that have a value of the mass_flame5 parameter that falls within our simulated primary mass range of (0.5 MMflame ≤ 1.5 M). The mass_flame provides an estimate of the stellar mass and comes from the FLAME module of the Astrophysical Parameters Inference System (Apsis, Creevey et al. 2023). For single stars this estimated mass refers to the mass of the star, and for binaries it corresponds to the mass of the primary object. Then, to generate synthetic stellar parameters for each star (either single or binary) we used a 2D kernel density estimation (KDE) approach (e.g., Silverman 1986; Scott 2015). This allowed us to conditionally sample values from a 2D KDE fit of the joint probability density function p(c, X) where c is the Gaia B PR P color of a source and X is a given DR3 parameter. Since our samples contain a combination of dwarf, sub-giants and giants, we considered that a non-parametric fitting approach such as KDE was more adequate to handle the potential multi-modality present in certain stellar parameters, and opted for kernel density estimation versus parametric distribution fitting. The average number of sources per 0.1 M mass bin in the range (0.6 MMflame ≤ 1.4 M) was in the order of thousands, see Figure A.1 (top panel), so this guaranteed that we had sufficient statistics for the KDE sampling. The generation of synthetic parameters was done as follows:

  1. We first divided each sample into mass bins of size 0.1 M using the parameter mass_flame.

  2. Then, for each star in the given sample and mass bin we extracted the pairs (ci, Xi), where ci is the bp_rp color and Xi is the value of the stellar parameter of interest (e.g., magnitude) for the i-th star in the sample.

  3. We then constructed a 2D kernel density estimator over the pairs (ci, Xi) to estimate the joint probability density function p(c, X): p^(c,X)=1Ni=1NK((c-ci,X-Xi);H),$\hat{p}(c, X)=\frac{1}{N} \sum_{i=1}^{N} K\left(\left(c-c_{i}, X-X_{i}\right) ; H\right),$(A.1) where K(x; H) is a Gaussian bivariate kernel and H is a socalled bandwidth matrix. The bandwidth matrix H controls the smoothness of the KDE, and is defined as: H=[hc2hcXhcXhX2],hjσj,j{c,X},$H=\left[\begin{array}{cc} h_{c}^{2} & h_{c X}\\ h_{c X} & h_{X}^{2} \end{array}\right], \quad h_{j} \propto \sigma_{j}, \quad j \in\{c, X\},$(A.2) where hj is proportional to the standard deviation of the corresponding variable σj and the off-diagonal elements hc X capture the correlation between color, c, and the variable, X. In practice K(x; H) is automatically computed using gaussian_kde from scipy.stats. 4. Now, for any desired color, c*, we compute the conditional probability distribution of X as p^(Xc*)=p^(c*,X)p^(c*,X)dX$\hat{p}\left(X \mid c^{*}\right)=\frac{\hat{p}\left(c^{*}, X\right)}{\int \hat{p}\left(c^{*}, X\right) d X}$(A.3) which yields a smooth, continuous estimate of p(Xc) for any value of c.

This procedure generated a library of M × N joint probability density functions p(Xc) where M is the number of mass bins and N is the number of synthetic parameters. Still, to use these joint PDFs, we needed to generate the missing value of the Gaia BP-RP color parameter for a given star with simulated mass M*. To do this we used the isochrone software package (Morton 2015) and performed a basic 1D interpolation using a main-sequence MIST isochrone (Bressan et al. 2012). We assumed solar metallicity and fixed 4.6GYr age for every star in our simulated samples, even though we now this is only roughly valid for a subset of the stars we used to generate the 2D KDEs library. However, given that: (1) the influence of photometry and spectroscopy parameters in the model decision is of second-order importance compared to the astrometric fit statistics, and (2) preserving correlations between parameters is of higher importance in this scenario than a high intrinsic parameter accuracy, we opted for this overall process of synthetic parameters generation versus one based on full usage of physical stellar models. Since the primary goal was to generate a plausible set of photometric and spectroscopic parameters, we considered that this approach was sufficient.

Nevertheless, to avoid any possible unphysical parameter combinations in our generated synthetic photometry, we implemented a coherence test that evaluates whether a simulated set of Gaia (BPRP) color, absolute Gaia magnitude (MG), and mass (M*) values are consistent with the physical properties of a single main-sequence star. To do this, we estimated the absolute G-band magnitude of a star with mass, M*, and color, B PR P, using the following expressions, MG pred (M,BP-RP)=M bol (M)-BCG(BP-RP),$M_{G}^{\mathrm{pred}}\left(M_{\star}, B P-R P\right)=M_{\mathrm{bol}}\left(M_{\star}\right)-B C_{G}(B P-R P),$(A.4) where Mbol is the bolometric magnitude associated with a simulated stellar mass M*, and B CG is the bolometric correction in the Gaia G band. The bolometric magnitude of the simulated star was computed as M bol (M)=M bol ,-2.5log10(L(M)L)$M_{\mathrm{bol}}\left(M_{\star}\right)=M_{\mathrm{bol}, \odot}-2.5 \log _{10}\left(\frac{L\left(M_{\star}\right)}{L_{\odot}}\right)$(A.5) where Mbol, ⊙ ≃ 4.74 and L(M)M4.0$L\left(M_{\star}\right) \simeq M_{\star}^{4.0}$ (Eker et al. 2015). The Gaia G-band bolometric correction (B CG) was determined by interpolating the empirical calibration grid presented in (Eker et al. 2020, Table 6). In this work the authors provide empirically derived bolometric corrections as a function of intrinsic stellar parameters for solar-type main-sequence stars. We then checked the difference between the predicted absolute magnitude for a given star and the synthetically generated one and only retained those synthetic values for which this difference was lower than a threshold: |MGpred-MGsynth|<τ=2σ$\left|M_{G}^{\text {pred }}-M_{G}^{\text {synth }}\right|<\tau=2 \sigma$ mag, where σ is the standard deviation of the computed residuals. We show in Figure A.2 the distribution of residuals MGpred-MGsynth$M_{G}^{\text {pred }}-M_{G}^{\text {synth }}$ and a comparison of the synthetic color-magnitude diagram obtained from our 2D KDE sampling procedure for both single and binary samples.

Now, while our selection procedure for single stars is designed to produce a sample of astrometrically well behaved (this is single) stars, this sample could still be contaminated with short-period binaries, near-equal-mass binaries with faceon orbits, or long-period binaries with a longer period than the DR3 time baseline. To assess the possible impact of introducing wrongly labelled binaries as single systems, we consider the f 1-score metric definition, f1=2·Precision·RecallPrecision+Recall= TP TP +12( FP + FN ),$f 1=\frac{2 \cdot \text { Precision } \cdot \text { Recall }}{\text { Precision }+ \text { Recall }}=\frac{\mathrm{TP}}{\mathrm{TP}+\frac{1}{2}(\mathrm{FP}+\mathrm{FN})},$(A.6) where TP refers to true positives, FP to false positives, and FN to false negatives. Since mislabelling binaries as single stars effectively increases FN, the f 1-score is expected to decrease as a consequence. We can analyse how a small perturbation Δ FN will affect the sensitivity of the f 1-score. If we hold TP and FP fixed and take the partial derivative with respect to FN: f1 FN =-12 TP ( TP +12( FP + FN ))2$\frac{\partial f 1}{\partial \mathrm{FN}}=-\frac{1}{2} \frac{\mathrm{TP}}{\left(\mathrm{TP}+\frac{1}{2}(\mathrm{FP}+\mathrm{FN})\right)^{2}}$(A.7) Hence, for a small change Δ FN, Δf112TP(TP+12(FP+FN))2ΔFNΔf1f1ΔFN2(TP+12(FP+FN))$\begin{align*} & \Delta f 1 \approx-\frac{1}{2} \frac{\mathrm{TP}}{\left(\mathrm{TP}+\frac{1}{2}(\mathrm{FP}+\mathrm{FN})\right)^{2}} \Delta \mathrm{FN}\\ & \frac{\Delta f 1}{f 1} \approx-\frac{\Delta \mathrm{FN}}{2\left(\mathrm{TP}+\frac{1}{2}(\mathrm{FP}+\mathrm{FN})\right)} \end{align*}$(A.8)

The magnitude of the drop in f 1-score depends mainly on the class balance. Using the derived formulae, we can analyse the expected degradation of performance for a balanced, strong model such as our ExoDNN, when a small fraction of FN is added. For example, assuming TP=900, FP=100, FN=100, and using equation A.6 we compute the f 1-score as: f1=900900+12(100+100)=0.9.$f 1=\frac{900}{900+\frac{1}{2}(100+100)}=0.9.$ The sensitivity in this score is then given by Equation A.7: f1 FN =-1290010002=-0.00045.$\frac{\partial f 1}{\partial \mathrm{FN}}=-\frac{1}{2} \frac{900}{1000^{2}}=-0.00045.$ Therefore, if we increase Δ FN=10 (i.e., +10%), using Equation A.8 we have: Δf1-0.00045×10=-0.0045,f1'0.8955,$\Delta f 1 \approx-0.00045 \times 10=-0.0045, \quad f 1^{\prime} \approx 0.8955,$ and we expect an absolute drop of 0.0045 in the f 1-score in this case. For Δ FN=5, the drop is halved, f 1 ≈ 0.8978. We can see how for a balanced model the impact of introducing a small fraction (≤ 10%) of false negatives is relatively small.

thumbnail Fig. A.2

TOP: Distribution of the residuals MGpredMGsynth$M_{G}^{{pred}}-M_{G}^{{synth}}$ for single and binary samples with corresponding Gaussian fits to the residuals. BOTTOM LEFT: Colour-magnitude diagram of the 100000 single (blue) sources and synthetic color-magnitude diagram with conditionally sampled color and magnitude values (orange). BOTTOM RIGHT: equivalent plot for the binaries with real binaries in (blue) and synthetic color-magnitude diagram with conditionally sampled color and magnitude values (orange).

Appendix B Additional models

For a relative performance comparison against ExoDNN, the selected binary classifiers were logistic regression and random forest. The logistic regression is similar to our network in the sense that is a probabilistic classifier whose objective is to minimize the log-loss, but it is fundamentally a linear model which cannot capture complex non-linear relationships unlike our ExoDNN. Random Forest, on the other side, is a non-parametric approach that uses an ensemble of decision trees to learn decision boundaries via recursive feature splits. We trained both classifiers using the same training data used to train ExoDNN and applied equivalent pre-processing steps, namely, robust scaling of the data before training. Table B.1 and Table B.2 list the different hyper-parameters used to train the alternative classifiers.

Table B.1

Best hyper-parameters for logistic regression

Table B.2

Best hyper-parameters for random forest

Appendix C Model input selection

To generate the input for the deep learning model prediction we selected sources from the Gaia main source table (gaiadr3.gaia_source) in the Gaia archive via the following query:

SELECT A.*, B.* FROM gaiadr3.gaia_source AS A JOIN gaiadr3.astrophysical_parameters AS B USING (source_id)
WHERE A.parallax > 10
AND A.visibility_periods_used>11
AND A.teff_gspphot > 2500
AND A.teff_gspphot < 7500
AND A.phot_variable_flag NOT LIKE ‘VARIABLE’

The join with the astrophysical parameters DR3 table (gaiadr3.astrophysical_parameters) is required to obtain estimated stellar parameters for each source generated by the astrophysical parameters inference system (Apsis, Creevey et al. 2023). Figure C.1 shows the number of sources retrieved by the query with a breakdown per spectral type.

thumbnail Fig. C.1

Comparison of the number of sources per spectral type between the inputs used to make model prediction and the FGKM golden sample (Creevey et al. 2023)

All Tables

Table 1

AGIS astrometric fit quality statistics.

Table 2

Gaia DR3 parameters used to train the model.

Table 3

Neural network hyper-parameters.

Table 4

Model evaluation results.

Table 5

Model testing results.

Table 6

ExoDNN closest-candidates prediction.

Table 7

ExoDNN prediction on external candidates.

Table B.1

Best hyper-parameters for logistic regression

Table B.2

Best hyper-parameters for random forest

All Figures

thumbnail Fig. 1

Left: simulated orbital periods (black) compared to the orbital periods from DR3 NSS Orbital solutions (shade), with the DR3 time baseline marked in vertical. Right: simulated secondary masses (black), compared to the catalogue of DR3 binary masses (shade), with the peak of the distribution of simulated secondary masses in vertical.

In the text
thumbnail Fig. 2

Top: Barycentre motion of a simulated single star corresponding to Eq. (2) with best fit astrometric parameters. The observed position of the source is marked with solid white circles, and the 1D along-scan observations with gray circles. Bottom: same, but for a simulated binary system, where the observed position of the source is perturbed by a companion (exaggerated for illustration purposes).

In the text
thumbnail Fig. 3

Left: evolution of the training and validation set losses over the different epochs. Right: evolution of the Brier-score during the training.

In the text
thumbnail Fig. 4

Left: classification threshold computation using the area under curve. Right: classification threshold computation using the precision-recall curve. The optimal threshold that achieves the best compromise between false positive and true positive rates is marked as a red dot. This is our chosen p0=0.242.

In the text
thumbnail Fig. 5

Top: parameter impact on the model output as estimated by the neural network model ranked by decreasing impact. High parameter values are color-coded in red, while low values are displayed in blue. Bottom: predicted probability for each binary star example on the test dataset (black dots), together with a detection probability proxy computed as p=1−Φ(S/N, δ)), where Φ is the CDF of the normal distribution, S/N=α/σ, and δ is a variable S/N threshold from 1 to 3.

In the text
thumbnail Fig. 6

Astrometric signature vs. orbital period for the simulated binary sources (orange), the 384 DR3 NSS ORBT solutions (magenta), the 1332 DR3 NSS ASB1 solutions (blue) and the currently confirmed exoplanets (Schneider et al. 2011) (gray). The red dashed vertical line represents the DR3 time baseline of 34 months in comparison to the shorter periods of the Gaia DR3 unresolved binary orbital solutions. For the DR3 NSS solutions we used the expression α=a0/d to compute the astrometric signature. a0 was computed using the Thiele-Innes elements (A, B, F, G) of the corresponding Gaia DR3 NSS solution and the recipe available in Halbwachs et al. (2023, Appendix A), and d was computed as the inverse of the source parallax. For the simulated sources we used α=M2/M1 × a1/d from Perryman et al. (2014), where M1 and M2 are the primary and secondary masses, respectively, and a1 is the semimajor axis of the orbit of the secondary.

In the text
thumbnail Fig. 7

Gaia B P minus R P color vs. absolute magnitude for the model input stars (green region). The FGKM golden sample is shown in yellow. Absolute magnitude is computed as: mG = phot_g_mean_mag + 5.0 · log(parallax)−10. Close binary systems where the smaller companion is a brown dwarf or a giant planet are displayed as diamonds (blue) and NSS Orbital* solutions are displayed as squares (magenta). CARMENES sources are displayed as circles (orange). Sources in correspond to the Gaia Catalog of Nearby Stars within d<100 pc.

In the text
thumbnail Fig. 8

Left: astrometric_excess_noise vs. Gaia color bp_rp. Model inputs are displayed in black and candidates in green. The vertical red lines show the applied color cuts. Orange line corresponds to the running median of the excess noise and the shaded orange region corresponds to the 16th-84th percentiles. Middle: equivalent plot but for ruwe vs. Gaia G band magnitude. The vertical red lines show the applied bright and faint Gaia magnitude cuts. Right: equivalent plot but for ipd_frac_multi_peak vs. Gaia G magnitude. The horizontal red line corresponds to the selected cut-off limit for this parameter.

In the text
thumbnail Fig. 9

Gaia BP minus RP color vs. absolute magnitude diagram of Gaia DR3 Catalog of Nearby Sources. The 7414 final candidates reported by ExoDNN after filtering are displayed in green. Prediction for a few reference objects, Proxima Centauri, LHS1610, BD+05 5128, HD 164604, Gaia-3b, Gaia-4b and Gaia-5b, etc., is also shown. Sources with known companions detected by ExoDNN are marked by the green boxes, while red-crosses refer to reference objects for which model predicted probability <p0, that is, undetected. ExoDNN closest new candidates are highlighted to the left of the main sequence with blue circles.

In the text
thumbnail Fig. 10

Top: distribution of the right ascension values for candidates sources (solid black line) and the original input for the model (gray) Middle: equivalent display but for Gaia BP minus RP color (bp-rp) distribution. Bottom: equivalent display but for the Gaia magnitude (phot_g_mean_mag) in this case.

In the text
thumbnail Fig. 11

Overall number of candidates per spectral type (spectraltype_esphs), pre-filtering (blue) and post-filtering (orange). Blue text labels on top of each group of bars correspond respectively to (s): median values of the astrometric_excess_ noise_sig parameter for the given spectral type and (f): median values of the ipd_frac_multi_peak parameter for the given spectral type.

In the text
thumbnail Fig. 12

Top: primary mass distribution of the candidate stars reported by ExoDNN (green). Also added for comparison are the mass distributions corresponding to the NSS Orbital* (orange), AstroSpectroSB1 (blue), and Eclipsing* binary (black) solutions from the DR3 NSS. The ExoDNN predictions show the 6% model false positive rate as downward arrows. Bottom: equivalent distributions but for metallicity of the candidates reported by ExoDNN.

In the text
thumbnail Fig. 13

ExoDNN G-type star candidates overlay in PLATO first field of view (LOPS1).

In the text
thumbnail Fig. A.1

TOP: Mass distribution of single (blue) and binary (orange) samples with bins of 0.1 M on the variable mass_flame. MIDDLE: Distribution of color indices for single sources (gray) compared to sampled colors (orange), obtained from a 1D KDE fit corresponding to the low end of the simulation mass range. BOTTOM: Equivalent plot as in the middle but for the binary sources in this case.

In the text
thumbnail Fig. A.2

TOP: Distribution of the residuals MGpredMGsynth$M_{G}^{{pred}}-M_{G}^{{synth}}$ for single and binary samples with corresponding Gaussian fits to the residuals. BOTTOM LEFT: Colour-magnitude diagram of the 100000 single (blue) sources and synthetic color-magnitude diagram with conditionally sampled color and magnitude values (orange). BOTTOM RIGHT: equivalent plot for the binaries with real binaries in (blue) and synthetic color-magnitude diagram with conditionally sampled color and magnitude values (orange).

In the text
thumbnail Fig. C.1

Comparison of the number of sources per spectral type between the inputs used to make model prediction and the FGKM golden sample (Creevey et al. 2023)

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.