[webkit-dev] is WebKit leaking or caching memory?

Luka Napotnik luka.napotnik at gmail.com
Mon Dec 1 04:22:08 PST 2008


I'm working on an application that uses WebKit as a "rendering farm". A
running program has a shared memory and an WebKit instance using the Gtk
+ port.

I've modified WebKit and added:
 - support for off-screen rendering, meaning instead of showing pages to
the screen I render them onto a cairo surface using the patches from the
Clutter project,
 - a function that simulates a click on a page with given X,Y
 - a new signal to the Gtk+ port that triggers whenever a visible change
occurred on the page (opening html, JavaScript modifications)

The HTML page has a DIV tag that is sensitive to mouse clicks. A click
on that DIV tag will cause the page to modify.

Now I have a client program that connects to the WebKit program using
IPC. Then I load the mentioned HTML page and sends click requests every
0.25 seconds with the coordinates where the DIV tag is located.

What I've notices is, that after a while, memory of the WebKit program
increases. The program was running for about one hour and the process
memory increased for about 7 MB.

Now I'm asking if the memory increment is a result of WebKit caching
subsystem or could that be a leak? Keep in mind that I click/refresh the
same page over and over again.


More information about the webkit-dev mailing list