How to create and manipulate Terabyte size Arrays with Win32API

How to create and manipulate Terabyte size Arrays with Win32API

  • Comments 1

If you are looking for a way of creating and accessing very large arrays, arrays that can handle content in the order of Tera Bytes, then probably you might find the File Mapping techniques useful for that purpose.

File mapping is the association of a file's contents with a portion of the virtual address space of a process. It allows the process to work efficiently with a large data files without having to map the whole file into memory. Processes read from and write to the file view using pointers, just as they would with dynamically allocated memory. This improves efficiency because the file resides on disk, but the file view resides in memory, the page-in and page-out happening seamlessly behind the scenes.

In a typical Win32 environment, applications have 4 gigabyte (GB) of virtual address space available. The virtual address space is divided so that 2 GB is available to the application and the other 2 GB is available only to the system. The size of a file mapping object that is backed by a named file is limited by disk space. The size of a file view is limited to the largest available contiguous block of unreserved virtual memory. This is at most 2 GB minus the virtual memory already reserved by the process.

To create a huge array, create multiple temporary files, each of _MaxFileSize, _MaxFileSize being any convenient maximum file size limit such as 2GB, and access their content by mapping and unmapping portions of them to the main memory as required. Creating the temporary files and mapping/unmapping them can be taken care by the CreateFile, CreateFileMapping, MapViewOfFile, and UnmapViewOfFile API. A simple implementation of this will look like below:

#define _TWOGB	(0x80000000)
#define _64KB (0x00010000)
#define _64MB (0x04000000)

class HugeArray
{
public:
enum : unsigned long
{ _ElemSize = sizeof(BYTE),
_ViewSize = _64KB,
_MaxFileSize = _TWOGB,
_MaxNoOfCachedViews = 16,
_MaxNoOfElemsPerFile= (_MaxFileSize/_ElemSize),
_MaxNoOfViewsPerFile= (_MaxFileSize/_ViewSize) //=0x00008000
};
protected:
HANDLE* hFileMappings; // The File mappings ATL::CAtlFile* Files; // The Files that were mapped unsigned int nFileCount; // Number of Files BIGNUMBER _RequestedLen; // No.of Elements requested BIGNUMBER _AllocatedLen; // No.of Elements actually allocated ATL::CAtlMap<unsigned int, LPVOID> pCachedViews; //[FileIndex, FileView] map for few cached views std::vector <CBitVector&> bvFileViewCleared; // Keeps track of views that need be cleared upon mapping public: HugeArray() // Directory where to create the memory mapped files : _RequestedLen(0), _AllocatedLen(0), hFileMappings(0), Files(0), nFileCount(0) { } // _nCount: Number of Elements to allocate space for
// lpszDirPath: Directory where to create the memory mapped files
// nInitVal: Value that should be used to initialize the allocated space
HugeArray(const BIGNUMBER& _nCount, LPCTSTR lpszDirPath = _T("."), int nInitVal=0) : _RequestedLen(0), _AllocatedLen(0), hFileMappings(0), Files(0), nFileCount(0) { Allocate(_nCount, lpszDirPath, nInitVal); } ~HugeArray() { DeAllocate(); } // Allocates space for _nCount number of elements (each of _ElemSize)
//returns the error code in case of failures
DWORD Allocate(const BIGNUMBER& _nCount, LPCTSTR lpszDirPath = _T("."), int nInitVal=0) { // Make sure things are clean DeAllocate(); // Prepare the file name prefixes ATL::CString strDir(lpszDirPath), strPrefix; TCHAR TempFileName[MAX_PATH+32]; // Make sure the directory path is terminate with "\" if(strDir.Right(1) != _T("\\")) strDir += _T("\\"); // Compute the required number of files unsigned long lNumFiles = (_nCount/(unsigned long)_MaxNoOfElemsPerFile).GetNumberIntoLong() + 1; // Allocate space for the handles if(NULL == (Files = new ATL::CAtlFile[lNumFiles]) || NULL == (hFileMappings = new HANDLE[lNumFiles])) return E_OUTOFMEMORY; for(unsigned long l=0; l < lNumFiles; ++l) { // Get a unique name for the file strPrefix.Format(_T("%dHA"), l); if(!GetTempFileName(strDir, strPrefix, 0, TempFileName)) return ErrorMessage(); // Create the file if(FAILED(Files[l].Create(TempFileName, GENERIC_READ|GENERIC_WRITE|STANDARD_RIGHTS_ALL, FILE_SHARE_READ, OPEN_ALWAYS, FILE_ATTRIBUTE_TEMPORARY| FILE_ATTRIBUTE_NOT_CONTENT_INDEXED| FILE_FLAG_DELETE_ON_CLOSE|FILE_FLAG_SEQUENTIAL_SCAN))) return ErrorMessage(); // Set the file size to the maximum
if(FAILED(Files[l].SetSize((0x00000000ffffffff & _MaxFileSize))))
return ErrorMessage();
// Create the File Mapping if(NULL == (hFileMappings[l] = ::CreateFileMapping(Files[l], 0, PAGE_READWRITE, 0, 0, 0))) return ErrorMessage(); // Create the Clearance vector bvFileViewCleared.push_back(CBitVector()); } _RequestedLen = _AllocatedLen = _nCount; nFileCount = lNumFiles; return S_OK; } void DeAllocate() { // Unmap the cached views POSITION pos = pCachedViews.GetStartPosition(); while(pos != NULL) { LPVOID pViewBase = pCachedViews.GetNext(pos)->m_value; if(pViewBase != NULL) ::UnmapViewOfFile(pViewBase); } pCachedViews.RemoveAll(); // Close the File Mapping Handles for(unsigned int i = 0; i < nFileCount; ++i) { if(hFileMappings[i] != NULL) ::CloseHandle(hFileMappings[i]); } // Free the File Mapping Handles memory if(hFileMappings != NULL) delete[] hFileMappings; // Free the Files Memory (which automatically closes the Files too) if(Files != NULL) delete[] Files; // Clear the bitvector bvFileViewCleared.clear(); // Reset Values to NULL hFileMappings = NULL; Files = NULL; nFileCount = 0; _AllocatedLen = 0; _RequestedLen = 0; } inline BYTE* _GetAt(unsigned long lFileIndex, unsigned long nViewIndex, unsigned long nViewOffset) { _ASSERTE(lFileIndex < nFileCount); unsigned long nFileViewIndex = ((lFileIndex & 0x0000ffff) << 16) | (nViewIndex & 0x0000ffff); LPVOID pViewBase = NULL; if(pCachedViews.Lookup(nFileViewIndex, pViewBase) == false) // Cache Miss { if(pCachedViews.GetCount() >= _MaxNoOfCachedViews) // if Cache is full, make room to load the new map { POSITION pos = pCachedViews.GetStartPosition(); pViewBase = pCachedViews.GetAt(pos)->m_value; if(pViewBase != NULL) ::UnmapViewOfFile(pViewBase); pCachedViews.RemoveAtPos(pos); } // Map the View if(NULL == (pViewBase = ::MapViewOfFile(hFileMappings[lFileIndex], FILE_MAP_WRITE, 0, nViewIndex*_ViewSize, _ViewSize))) { ATL::CString strFilePath; strFilePath.Format(_T("File Index: %d"), lFileIndex); ::MessageBox(NULL, _T("Unable to map view for: ") + strFilePath, _T("Error"), MB_OK|MB_ICONERROR); return (BYTE*)pViewBase; } // Clear the contents of the view if this is the first time it is mapped if(bvFileViewCleared[lFileIndex].IsBitSet(nViewIndex)==false) { ::memset(pViewBase, 0, _ViewSize); bvFileViewCleared[lFileIndex] += nViewIndex; // set the cleared flag } // Cache the mapped view pCachedViews[nFileViewIndex] = pViewBase; } return ((BYTE*)pViewBase + nViewOffset); } inline BYTE& operator[](const BIGNUMBER& nIndex) { unsigned long lFileIndex = (nIndex / (unsigned long)_MaxNoOfElemsPerFile).GetNumberIntoLong(); unsigned long lFileOffset = (nIndex % (unsigned long)_MaxNoOfElemsPerFile).GetNumberIntoLong(); // Compute the offset in file unsigned long nViewIndex = (lFileOffset / (unsigned long)_ViewSize); // Compute the view that should be containing the offset unsigned long nViewOffset = (lFileOffset % (unsigned long)_ViewSize); // Compute the offset in the view return *_GetAt(lFileIndex, nViewIndex, nViewOffset); } inline BYTE& operator[](const unsigned long ulIndex) { unsigned long lFileIndex = (ulIndex / (unsigned long)_MaxNoOfElemsPerFile); unsigned long lFileOffset = (ulIndex % (unsigned long)_MaxNoOfElemsPerFile); // Compute the offset in file unsigned long nViewIndex = (lFileOffset / (unsigned long)_ViewSize); // Compute the view that should be containing the offset unsigned long nViewOffset = (lFileOffset % (unsigned long)_ViewSize); // Compute the offset in the view return *_GetAt(lFileIndex, nViewIndex, nViewOffset); } // Attemps to allocate space for _nCount number of elements.
// If the currently allocate space is more than what is required, this would do nothing.
// Else Deallocates the existing space and allocates new space.
// In any case clears the contents.
bool ReDim(const BIGNUMBER& _nCount, LPCTSTR lpszDirPath = _T("."), int nInitVal=0) { if(_AllocatedLen < _nCount) { return (S_OK == Allocate(_nCount, lpszDirPath, nInitVal)); } else { _RequestedLen = _nCount; ClearContents(); } return false; } void ClearContents() { // Unmap the cached views POSITION pos = pCachedViews.GetStartPosition(); while(pos != NULL) { LPVOID pViewBase = pCachedViews.GetNext(pos)->m_value; if(pViewBase != NULL) ::UnmapViewOfFile(pViewBase); } pCachedViews.RemoveAll(); // Clear the Clearance vector and Create it again bvFileViewCleared.clear(); for(unsigned int i = 0; i < nFileCount; ++i) bvFileViewCleared.push_back(CBitVector()); } const BIGNUMBER& Length() const { return _RequestedLen; } };

The HugeArray class shown above facilitates large Byte arrays to be allocated and accessed by using the file mapping. It uses multiple files of _MaxFileSize to allocate the space and maps views on the files. Views are mapped and unmapped as required, with a limited number of views cached at a time. _MaxNoOfCachedViews governs this limit.

CachedViews are picked in FIFO fashion to unmap so as to make space for new views. (FIFO could be replaced with LRU or any other efficent techqniue, but that requires keeping track of the view usage and hence might incur extra processing - but then the best method has to be decided based on the target scenario.)

The _ViewSize governs the size of the view that is mapped. (_MaxFileSize/_ViewSize) gives the max no. of views per file. As per our other metrics above, since the MaxNoOfViewsPerFile require only 16bits, we can use the upper 16 bits (of a 32-bit integer) to keep track of which file this view belongs to. Thus, we are restricting the maximum number of files to be 2^16 (=65536). With 2GB per file, this restricts us to have at most 65536 * 2GB address space. Thus our HugeArray class cannot work for arrays that require more than 131072GB space.

If we require more space to be mapped, then we need to maintain the FileIndex and View offset seperately (instead of merging them both into one 32-bit integer), thus increasing the number of bits for FileIndex making it possible to have more files mapped.

A typical usage of this call will look like:

#include "HugeArray.h"

int _tmain(int argc, _TCHAR* argv[])
{
BIGNUMBER bSize("4294967296"); // 4GB HugeArray ha(bSize, _T(".")); // Lets allocate an array of 4GB ha[0] = 0; ha[1] = 1; ha[2] = 2; ha[bSize-1] = (BYTE)((bSize-1).GetNumberIntoLong() % 256); ha[bSize-2] = (BYTE)((bSize-2).GetNumberIntoLong() % 256); for(BIGNUMBER i=bSize-3; i >= 0; --i) { ha[i] = (BYTE)(i.GetNumberIntoLong())%256; if(ha[i+2] != ((i+2).GetNumberIntoLong()%256)) printf("\nError at %s\n", i.GetNumberIntoString(sz)); } }

A sample project demonstrating this can be downloaded from the attachments.

Ofcourse, this class can be templatized to allow the elements to complex records than just Bytes. Only that care should be taken to ensure the records fall within the file boundary and do not spawn across.

A point worth noting is, this mechanism relies on the assumption that we never require to have all the content in the memory at once and that the element access follow the temporal and special proximity when possible.

Attachment: HugeArrays.zip
  • Debug Setting .NET breakpoints in Windbg for applications that crash on startup Web Javascript HTML Construction

Page 1 of 1 (1 items)
Leave a Comment
  • Please add 5 and 8 and type the answer here:
  • Post