Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Graphbolt][Performance] Reduce the memory usage of preprocess_ondisk_dataset #7086

Open
czkkkkkk opened this issue Feb 5, 2024 · 2 comments
Assignees

Comments

@czkkkkkk
Copy link
Collaborator

czkkkkkk commented Feb 5, 2024

🚀 Feature

Motivation

Currently, preprocess_ondisk_dataset consumes much more memory than the topology of a graph itself during the preprocessing. When loading a graph with 2B nodes and 8B edges, it cannot be finished in a machine with 380 GB memory. After a rough profiling, I found that the peak memory usage is reached when converting a DGL graph to a fused sampling graph.

fused_csc_sampling_graph = from_dglgraph(

There could be two factors contributing to the peak memory usage.

  1. The input DGL graph is passed to the function, which consumes about 160 GB memory.
  2. from_dglgraph creates a temporary homogeneous graph and also its CSC format.

Alternatives

Pitch

Additional context

@Rhett-Ying
Copy link
Collaborator

@Skeleton003 could you look into it and try with the new implementation: #6986

Copy link

github-actions bot commented Mar 7, 2024

This issue has been automatically marked as stale due to lack of activity. It will be closed if no further activity occurs. Thank you

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants