Jump to content

Recommended Posts


I am trying to model large memories (>8 GB) on a virtual platform that I am working on. I don't think using the C++ 'new' operator to allocate the entire chunk is a good idea. Can someone suggest any methods they think or have used in the past to model such? My memory is going to be very sparse to start with and might start filling up only at a later time.


Link to comment
Share on other sites

If very small memory locations are used in entire memory (which is >8GB) you can use on demand paging where the memory is allocated in terms of pages whenever the page is not allocated for requested address. This enables allocate only pages whenever required. You can achieve overall good simulation speed.

However in above mechanism if more and more pages allocated, the simulation speed will be downgraded or can crash also because of page allocation failure.

In this case, memory can be implemented in file and using IO operations it can be accessed. With this simulation performance will be degraded but ensures simulation will not stop because of page allocation failure.

Link to comment
Share on other sites

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.

Reply to this topic...

×   Pasted as rich text.   Paste as plain text instead

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

  • Create New...