[petsc-users] Accessing submatrices without additional memory usage
Michał Dereziński
michal.derezinski at gmail.com
Wed May 24 13:53:07 CDT 2017
It is an optimization problem minimizing a convex objective for a binary classification task, which I’m solving using a Tao solver.
The multiplication operations are performing gradient computation for each step of the optimization.
So I’m performing both a MatMult and a MatMultTranspose, in both cases the vector may be a dense vector.
The crucial part of the implementation is that at the beginning I am not running on the entire dataset (rows of the full matrix).
As a consequence I don’t need to have the entire matrix loaded right away. In fact, in some cases I may choose to stop the optimization before the entire matrix has been loaded (I already verified that this scenario may come up as a use case). That is why it is important that I don’t load it at the beginning.
Parallel loading is not a necessary part of the implementation. Initially, I intend to alternate between loading a portion of the matrix, then doing computations, then loading more of the matrix, etc. But, given that I observed large loading times for some datasets, parallel loading may make sense, if done efficiently.
Thanks,
Michal.
> Wiadomość napisana przez Jed Brown <jed at jedbrown.org> w dniu 24.05.2017, o godz. 11:32:
>
> Michał Dereziński <michal.derezinski at gmail.com> writes:
>
>> Great! Then I have a follow-up question:
>>
>> My goal is to be able to load the full matrix X from disk, while at
>> the same time in parallel, performing computations on the submatrices
>> that have already been loaded. Essentially, I want to think of X as a
>> block matrix (where the blocks are horizontal, spanning the full width
>> of the matrix),
>
> What would be the distribution of the vector that this non-square
> submatrix (probably with many empty columns) is applied to?
>
> Could you back up and explain what problem you're trying to solve? It
> sounds like you're about to code yourself into a dungeon.
More information about the petsc-users
mailing list