Several challenges emerged from the dataclysm of neuroimaging datasets spanning both healthy and disordered brain spectrum. In particular, samples with missing data views (e.g., functional imaging modality) constitute a hurdle to conventional big data learning techniques which ideally would be trained using a maximum number of samples across all views. Existing works on predicting target data views from a source data view mainly used brain images such as predicting PET image from MRI image. However, to the best of our knowledge, predicting a set of target brain networks from a source network remains unexplored. To fill this gap, a multi-kernel manifold learning (MKML) framework is proposed to learn how to predictmulti-view brain networks from a source network to impute missing views in a connectomic dataset. Prior to performing multiple kernel learning of multi-view data, it is typically assumed that the source and target data come from the same distribution. However, multi-view connectomic data can be drawn from different distributions. In order to build robust predictors for predicting target multi-view networks from a source network view, it is necessary to take into account the shift between the source and target domains. Hence, we first estimate a mapping function that transforms the source and the target domains into a shared space where their correlation is maximized using canonical correlation analysis (CCA). Next, we nest the projected training and testing source samples into a connectomic manifold using multiple kernel learning, where we identify the most similar training samples to the testing source network. Given a testing subject, we introduce a cross-domain trust score to assess the reliability of each selected training sample for the target prediction task. Our model outperformed both conventional MKML technique and the proposed CCA-based MKML technique without enhancement by trust scores.