I am integrating proteomics, peptidomics and physicochemical data using the DIABLO framework. I have been referencing your excellent 2019 publication on DIABLO, and I find the methodology very powerful for my research.
I had a question regarding block weighting when dealing with datasets of significantly different sizes. In my case, I am working with a proteomics dataset (~3600 variables) and a peptidomics dataset (~324 variables), along with a smaller block of physicochemical properties (only 3 variables). I am concerned about the disparity in variable numbers across these blocks and their influence on the integration and feature selection process.
In your paper, I didn’t find specific guidance on adjusting the weights of blocks in the design matrix to account for this imbalance. Could you kindly clarify if block weighting to adjust for the size of blocks (or other factors) is recommended in DIABLO? If so, could you point me to any literature or specific methodology regarding how to effectively implement such an approach in DIABLO?