Which graph library is best suited for large graphs?

geremy condra debatem1 at gmail.com
Fri Dec 11 21:48:34 EST 2009


On Fri, Dec 11, 2009 at 5:12 AM, Wolodja Wentland
<wentland at cl.uni-heidelberg.de> wrote:
> Hi all,
>
> I am writing a library for accessing Wikipedia data and include a module
> that generates graphs from the Link structure between articles and other
> pages (like categories).
>
> These graphs could easily contain some million nodes which are frequently
> linked. The graphs I am building right now have around 300.000 nodes
> with an average in/out degree of - say - 4 and already need around 1-2GB of
> memory. I use networkx to model the graphs and serialise them to files on
> the disk. (using adjacency list format, pickle and/or graphml).

Huh. Using graphine- which should be somewhat more memory hungry
than networkx- I generated a naive million node 4-cycle graph and wound
up using something under 600 meg of ram. Can you post some code?

Geremy Condra



More information about the Python-list mailing list