b486678290
Library -Artifacts
1458 lines
55 KiB
C#
1458 lines
55 KiB
C#
using System;
|
|
using System.Collections;
|
|
using System.Collections.Generic;
|
|
using System.Diagnostics;
|
|
using System.Runtime.InteropServices;
|
|
using System.Threading;
|
|
using Unity.Burst;
|
|
using Unity.Mathematics;
|
|
using Unity.Jobs;
|
|
using Unity.Jobs.LowLevel.Unsafe;
|
|
using UnityEngine.Assertions;
|
|
|
|
namespace Unity.Collections.LowLevel.Unsafe
|
|
{
|
|
/// <summary>
|
|
/// A bucket of key-value pairs. Used as the internal storage for hash maps.
|
|
/// </summary>
|
|
/// <remarks>Exposed publicly only for advanced use cases.</remarks>
|
|
[BurstCompatible]
|
|
public unsafe struct UnsafeHashMapBucketData
|
|
{
|
|
internal UnsafeHashMapBucketData(byte* v, byte* k, byte* n, byte* b, int bcm)
|
|
{
|
|
values = v;
|
|
keys = k;
|
|
next = n;
|
|
buckets = b;
|
|
bucketCapacityMask = bcm;
|
|
}
|
|
|
|
/// <summary>
|
|
/// The buffer of values.
|
|
/// </summary>
|
|
/// <value>The buffer of values.</value>
|
|
public readonly byte* values;
|
|
|
|
/// <summary>
|
|
/// The buffer of keys.
|
|
/// </summary>
|
|
/// <value>The buffer of keys.</value>
|
|
public readonly byte* keys;
|
|
|
|
/// <summary>
|
|
/// The next bucket in the chain.
|
|
/// </summary>
|
|
/// <value>The next bucket in the chain.</value>
|
|
public readonly byte* next;
|
|
|
|
/// <summary>
|
|
/// The first bucket in the chain.
|
|
/// </summary>
|
|
/// <value>The first bucket in the chain.</value>
|
|
public readonly byte* buckets;
|
|
|
|
/// <summary>
|
|
/// One less than the bucket capacity.
|
|
/// </summary>
|
|
/// <value>One less than the bucket capacity.</value>
|
|
public readonly int bucketCapacityMask;
|
|
}
|
|
|
|
[StructLayout(LayoutKind.Explicit)]
|
|
[BurstCompatible]
|
|
internal unsafe struct UnsafeParallelHashMapData
|
|
{
|
|
[FieldOffset(0)]
|
|
internal byte* values;
|
|
// 4-byte padding on 32-bit architectures here
|
|
|
|
[FieldOffset(8)]
|
|
internal byte* keys;
|
|
// 4-byte padding on 32-bit architectures here
|
|
|
|
[FieldOffset(16)]
|
|
internal byte* next;
|
|
// 4-byte padding on 32-bit architectures here
|
|
|
|
[FieldOffset(24)]
|
|
internal byte* buckets;
|
|
// 4-byte padding on 32-bit architectures here
|
|
|
|
[FieldOffset(32)]
|
|
internal int keyCapacity;
|
|
|
|
[FieldOffset(36)]
|
|
internal int bucketCapacityMask; // = bucket capacity - 1
|
|
|
|
[FieldOffset(40)]
|
|
internal int allocatedIndexLength;
|
|
|
|
[FieldOffset(JobsUtility.CacheLineSize < 64 ? 64 : JobsUtility.CacheLineSize)]
|
|
internal fixed int firstFreeTLS[JobsUtility.MaxJobThreadCount * IntsPerCacheLine];
|
|
|
|
// 64 is the cache line size on x86, arm usually has 32 - so it is possible to save some memory there
|
|
internal const int IntsPerCacheLine = JobsUtility.CacheLineSize / sizeof(int);
|
|
|
|
internal static int GetBucketSize(int capacity)
|
|
{
|
|
return capacity * 2;
|
|
}
|
|
|
|
internal static int GrowCapacity(int capacity)
|
|
{
|
|
if (capacity == 0)
|
|
{
|
|
return 1;
|
|
}
|
|
|
|
return capacity * 2;
|
|
}
|
|
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int), typeof(int) })]
|
|
internal static void AllocateHashMap<TKey, TValue>(int length, int bucketLength, AllocatorManager.AllocatorHandle label,
|
|
out UnsafeParallelHashMapData* outBuf)
|
|
where TKey : struct
|
|
where TValue : struct
|
|
{
|
|
CollectionHelper.CheckIsUnmanaged<TKey>();
|
|
CollectionHelper.CheckIsUnmanaged<TValue>();
|
|
|
|
UnsafeParallelHashMapData* data = (UnsafeParallelHashMapData*)Memory.Unmanaged.Allocate(sizeof(UnsafeParallelHashMapData), UnsafeUtility.AlignOf<UnsafeParallelHashMapData>(), label);
|
|
|
|
bucketLength = math.ceilpow2(bucketLength);
|
|
|
|
data->keyCapacity = length;
|
|
data->bucketCapacityMask = bucketLength - 1;
|
|
|
|
int keyOffset, nextOffset, bucketOffset;
|
|
int totalSize = CalculateDataSize<TKey, TValue>(length, bucketLength, out keyOffset, out nextOffset, out bucketOffset);
|
|
|
|
data->values = (byte*)Memory.Unmanaged.Allocate(totalSize, JobsUtility.CacheLineSize, label);
|
|
data->keys = data->values + keyOffset;
|
|
data->next = data->values + nextOffset;
|
|
data->buckets = data->values + bucketOffset;
|
|
|
|
outBuf = data;
|
|
}
|
|
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int), typeof(int) })]
|
|
internal static void ReallocateHashMap<TKey, TValue>(UnsafeParallelHashMapData* data, int newCapacity, int newBucketCapacity, AllocatorManager.AllocatorHandle label)
|
|
where TKey : struct
|
|
where TValue : struct
|
|
{
|
|
newBucketCapacity = math.ceilpow2(newBucketCapacity);
|
|
|
|
if (data->keyCapacity == newCapacity && (data->bucketCapacityMask + 1) == newBucketCapacity)
|
|
{
|
|
return;
|
|
}
|
|
|
|
CheckHashMapReallocateDoesNotShrink(data, newCapacity);
|
|
|
|
int keyOffset, nextOffset, bucketOffset;
|
|
int totalSize = CalculateDataSize<TKey, TValue>(newCapacity, newBucketCapacity, out keyOffset, out nextOffset, out bucketOffset);
|
|
|
|
byte* newData = (byte*)Memory.Unmanaged.Allocate(totalSize, JobsUtility.CacheLineSize, label);
|
|
byte* newKeys = newData + keyOffset;
|
|
byte* newNext = newData + nextOffset;
|
|
byte* newBuckets = newData + bucketOffset;
|
|
|
|
// The items are taken from a free-list and might not be tightly packed, copy all of the old capcity
|
|
UnsafeUtility.MemCpy(newData, data->values, data->keyCapacity * UnsafeUtility.SizeOf<TValue>());
|
|
UnsafeUtility.MemCpy(newKeys, data->keys, data->keyCapacity * UnsafeUtility.SizeOf<TKey>());
|
|
UnsafeUtility.MemCpy(newNext, data->next, data->keyCapacity * UnsafeUtility.SizeOf<int>());
|
|
|
|
for (int emptyNext = data->keyCapacity; emptyNext < newCapacity; ++emptyNext)
|
|
{
|
|
((int*)newNext)[emptyNext] = -1;
|
|
}
|
|
|
|
// re-hash the buckets, first clear the new bucket list, then insert all values from the old list
|
|
for (int bucket = 0; bucket < newBucketCapacity; ++bucket)
|
|
{
|
|
((int*)newBuckets)[bucket] = -1;
|
|
}
|
|
|
|
for (int bucket = 0; bucket <= data->bucketCapacityMask; ++bucket)
|
|
{
|
|
int* buckets = (int*)data->buckets;
|
|
int* nextPtrs = (int*)newNext;
|
|
while (buckets[bucket] >= 0)
|
|
{
|
|
int curEntry = buckets[bucket];
|
|
buckets[bucket] = nextPtrs[curEntry];
|
|
int newBucket = UnsafeUtility.ReadArrayElement<TKey>(data->keys, curEntry).GetHashCode() & (newBucketCapacity - 1);
|
|
nextPtrs[curEntry] = ((int*)newBuckets)[newBucket];
|
|
((int*)newBuckets)[newBucket] = curEntry;
|
|
}
|
|
}
|
|
|
|
Memory.Unmanaged.Free(data->values, label);
|
|
if (data->allocatedIndexLength > data->keyCapacity)
|
|
{
|
|
data->allocatedIndexLength = data->keyCapacity;
|
|
}
|
|
|
|
data->values = newData;
|
|
data->keys = newKeys;
|
|
data->next = newNext;
|
|
data->buckets = newBuckets;
|
|
data->keyCapacity = newCapacity;
|
|
data->bucketCapacityMask = newBucketCapacity - 1;
|
|
}
|
|
|
|
internal static void DeallocateHashMap(UnsafeParallelHashMapData* data, AllocatorManager.AllocatorHandle allocator)
|
|
{
|
|
Memory.Unmanaged.Free(data->values, allocator);
|
|
Memory.Unmanaged.Free(data, allocator);
|
|
}
|
|
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int), typeof(int) })]
|
|
internal static int CalculateDataSize<TKey, TValue>(int length, int bucketLength, out int keyOffset, out int nextOffset, out int bucketOffset)
|
|
where TKey : struct
|
|
where TValue : struct
|
|
{
|
|
var sizeOfTValue = UnsafeUtility.SizeOf<TValue>();
|
|
var sizeOfTKey = UnsafeUtility.SizeOf<TKey>();
|
|
var sizeOfInt = UnsafeUtility.SizeOf<int>();
|
|
|
|
var valuesSize = CollectionHelper.Align(sizeOfTValue * length, JobsUtility.CacheLineSize);
|
|
var keysSize = CollectionHelper.Align(sizeOfTKey * length, JobsUtility.CacheLineSize);
|
|
var nextSize = CollectionHelper.Align(sizeOfInt * length, JobsUtility.CacheLineSize);
|
|
var bucketSize = CollectionHelper.Align(sizeOfInt * bucketLength, JobsUtility.CacheLineSize);
|
|
var totalSize = valuesSize + keysSize + nextSize + bucketSize;
|
|
|
|
keyOffset = 0 + valuesSize;
|
|
nextOffset = keyOffset + keysSize;
|
|
bucketOffset = nextOffset + nextSize;
|
|
|
|
return totalSize;
|
|
}
|
|
|
|
internal static bool IsEmpty(UnsafeParallelHashMapData* data)
|
|
{
|
|
if (data->allocatedIndexLength <= 0)
|
|
{
|
|
return true;
|
|
}
|
|
|
|
var bucketArray = (int*)data->buckets;
|
|
var bucketNext = (int*)data->next;
|
|
var capacityMask = data->bucketCapacityMask;
|
|
|
|
for (int i = 0; i <= capacityMask; ++i)
|
|
{
|
|
int bucket = bucketArray[i];
|
|
|
|
if (bucket != -1)
|
|
{
|
|
return false;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
internal static int GetCount(UnsafeParallelHashMapData* data)
|
|
{
|
|
if (data->allocatedIndexLength <= 0)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
var bucketNext = (int*)data->next;
|
|
var freeListSize = 0;
|
|
|
|
for (int tls = 0; tls < JobsUtility.MaxJobThreadCount; ++tls)
|
|
{
|
|
for (var freeIdx = data->firstFreeTLS[tls * IntsPerCacheLine]
|
|
; freeIdx >= 0
|
|
; freeIdx = bucketNext[freeIdx]
|
|
)
|
|
{
|
|
++freeListSize;
|
|
}
|
|
}
|
|
|
|
return math.min(data->keyCapacity, data->allocatedIndexLength) - freeListSize;
|
|
}
|
|
|
|
internal static bool MoveNext(UnsafeParallelHashMapData* data, ref int bucketIndex, ref int nextIndex, out int index)
|
|
{
|
|
var bucketArray = (int*)data->buckets;
|
|
var bucketNext = (int*)data->next;
|
|
var capacityMask = data->bucketCapacityMask;
|
|
|
|
if (nextIndex != -1)
|
|
{
|
|
index = nextIndex;
|
|
nextIndex = bucketNext[nextIndex];
|
|
return true;
|
|
}
|
|
|
|
for (int i = bucketIndex; i <= capacityMask; ++i)
|
|
{
|
|
var idx = bucketArray[i];
|
|
|
|
if (idx != -1)
|
|
{
|
|
index = idx;
|
|
bucketIndex = i + 1;
|
|
nextIndex = bucketNext[idx];
|
|
|
|
return true;
|
|
}
|
|
}
|
|
|
|
index = -1;
|
|
bucketIndex = capacityMask + 1;
|
|
nextIndex = -1;
|
|
return false;
|
|
}
|
|
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int) })]
|
|
internal static void GetKeyArray<TKey>(UnsafeParallelHashMapData* data, NativeArray<TKey> result)
|
|
where TKey : struct
|
|
{
|
|
var bucketArray = (int*)data->buckets;
|
|
var bucketNext = (int*)data->next;
|
|
|
|
for (int i = 0, count = 0, max = result.Length; i <= data->bucketCapacityMask && count < max; ++i)
|
|
{
|
|
int bucket = bucketArray[i];
|
|
|
|
while (bucket != -1)
|
|
{
|
|
result[count++] = UnsafeUtility.ReadArrayElement<TKey>(data->keys, bucket);
|
|
bucket = bucketNext[bucket];
|
|
}
|
|
}
|
|
}
|
|
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int) })]
|
|
internal static void GetValueArray<TValue>(UnsafeParallelHashMapData* data, NativeArray<TValue> result)
|
|
where TValue : struct
|
|
{
|
|
var bucketArray = (int*)data->buckets;
|
|
var bucketNext = (int*)data->next;
|
|
|
|
for (int i = 0, count = 0, max = result.Length, capacityMask = data->bucketCapacityMask
|
|
; i <= capacityMask && count < max
|
|
; ++i
|
|
)
|
|
{
|
|
int bucket = bucketArray[i];
|
|
|
|
while (bucket != -1)
|
|
{
|
|
result[count++] = UnsafeUtility.ReadArrayElement<TValue>(data->values, bucket);
|
|
bucket = bucketNext[bucket];
|
|
}
|
|
}
|
|
}
|
|
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int), typeof(int) })]
|
|
internal static void GetKeyValueArrays<TKey, TValue>(UnsafeParallelHashMapData* data, NativeKeyValueArrays<TKey, TValue> result)
|
|
where TKey : struct
|
|
where TValue : struct
|
|
{
|
|
var bucketArray = (int*)data->buckets;
|
|
var bucketNext = (int*)data->next;
|
|
|
|
for (int i = 0, count = 0, max = result.Length, capacityMask = data->bucketCapacityMask
|
|
; i <= capacityMask && count < max
|
|
; ++i
|
|
)
|
|
{
|
|
int bucket = bucketArray[i];
|
|
|
|
while (bucket != -1)
|
|
{
|
|
result.Keys[count] = UnsafeUtility.ReadArrayElement<TKey>(data->keys, bucket);
|
|
result.Values[count] = UnsafeUtility.ReadArrayElement<TValue>(data->values, bucket);
|
|
count++;
|
|
bucket = bucketNext[bucket];
|
|
}
|
|
}
|
|
}
|
|
|
|
internal UnsafeHashMapBucketData GetBucketData()
|
|
{
|
|
return new UnsafeHashMapBucketData(values, keys, next, buckets, bucketCapacityMask);
|
|
}
|
|
|
|
[Conditional("ENABLE_UNITY_COLLECTIONS_CHECKS")]
|
|
static void CheckHashMapReallocateDoesNotShrink(UnsafeParallelHashMapData* data, int newCapacity)
|
|
{
|
|
if (data->keyCapacity > newCapacity)
|
|
throw new Exception("Shrinking a hash map is not supported");
|
|
}
|
|
}
|
|
|
|
[NativeContainer]
|
|
[BurstCompatible]
|
|
internal unsafe struct UnsafeParallelHashMapDataDispose
|
|
{
|
|
[NativeDisableUnsafePtrRestriction]
|
|
internal UnsafeParallelHashMapData* m_Buffer;
|
|
internal AllocatorManager.AllocatorHandle m_AllocatorLabel;
|
|
|
|
#if ENABLE_UNITY_COLLECTIONS_CHECKS
|
|
internal AtomicSafetyHandle m_Safety;
|
|
#endif
|
|
|
|
public void Dispose()
|
|
{
|
|
UnsafeParallelHashMapData.DeallocateHashMap(m_Buffer, m_AllocatorLabel);
|
|
}
|
|
}
|
|
|
|
[BurstCompile]
|
|
internal unsafe struct UnsafeParallelHashMapDataDisposeJob : IJob
|
|
{
|
|
internal UnsafeParallelHashMapDataDispose Data;
|
|
|
|
public void Execute()
|
|
{
|
|
Data.Dispose();
|
|
}
|
|
}
|
|
|
|
[StructLayout(LayoutKind.Sequential)]
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int), typeof(int) })]
|
|
internal struct UnsafeParallelHashMapBase<TKey, TValue>
|
|
where TKey : struct, IEquatable<TKey>
|
|
where TValue : struct
|
|
{
|
|
internal static unsafe void Clear(UnsafeParallelHashMapData* data)
|
|
{
|
|
UnsafeUtility.MemSet(data->buckets, 0xff, (data->bucketCapacityMask + 1) * 4);
|
|
UnsafeUtility.MemSet(data->next, 0xff, (data->keyCapacity) * 4);
|
|
|
|
for (int tls = 0; tls < JobsUtility.MaxJobThreadCount; ++tls)
|
|
{
|
|
data->firstFreeTLS[tls * UnsafeParallelHashMapData.IntsPerCacheLine] = -1;
|
|
}
|
|
|
|
data->allocatedIndexLength = 0;
|
|
}
|
|
|
|
internal static unsafe int AllocEntry(UnsafeParallelHashMapData* data, int threadIndex)
|
|
{
|
|
int idx;
|
|
int* nextPtrs = (int*)data->next;
|
|
|
|
do
|
|
{
|
|
idx = data->firstFreeTLS[threadIndex * UnsafeParallelHashMapData.IntsPerCacheLine];
|
|
|
|
// Check if this thread has a free entry. Negative value means there is nothing free.
|
|
if (idx < 0)
|
|
{
|
|
// Try to refill local cache. The local cache is a linked list of 16 free entries.
|
|
|
|
// Indicate to other threads that we are refilling the cache.
|
|
// -2 means refilling cache.
|
|
// -1 means nothing free on this thread.
|
|
Interlocked.Exchange(ref data->firstFreeTLS[threadIndex * UnsafeParallelHashMapData.IntsPerCacheLine], -2);
|
|
|
|
// If it failed try to get one from the never-allocated array
|
|
if (data->allocatedIndexLength < data->keyCapacity)
|
|
{
|
|
idx = Interlocked.Add(ref data->allocatedIndexLength, 16) - 16;
|
|
|
|
if (idx < data->keyCapacity - 1)
|
|
{
|
|
int count = math.min(16, data->keyCapacity - idx);
|
|
|
|
// Set up a linked list of free entries.
|
|
for (int i = 1; i < count; ++i)
|
|
{
|
|
nextPtrs[idx + i] = idx + i + 1;
|
|
}
|
|
|
|
// Last entry points to null.
|
|
nextPtrs[idx + count - 1] = -1;
|
|
|
|
// The first entry is going to be allocated to someone so it also points to null.
|
|
nextPtrs[idx] = -1;
|
|
|
|
// Set the TLS first free to the head of the list, which is the one after the entry we are returning.
|
|
Interlocked.Exchange(ref data->firstFreeTLS[threadIndex * UnsafeParallelHashMapData.IntsPerCacheLine], idx + 1);
|
|
|
|
return idx;
|
|
}
|
|
|
|
if (idx == data->keyCapacity - 1)
|
|
{
|
|
// We tried to allocate more entries for this thread but we've already hit the key capacity,
|
|
// so we are in fact out of space. Record that this thread has no more entries.
|
|
Interlocked.Exchange(ref data->firstFreeTLS[threadIndex * UnsafeParallelHashMapData.IntsPerCacheLine], -1);
|
|
|
|
return idx;
|
|
}
|
|
}
|
|
|
|
// If we reach here, then we couldn't allocate more entries for this thread, so it's completely empty.
|
|
Interlocked.Exchange(ref data->firstFreeTLS[threadIndex * UnsafeParallelHashMapData.IntsPerCacheLine], -1);
|
|
|
|
// Failed to get any, try to get one from another free list
|
|
bool again = true;
|
|
while (again)
|
|
{
|
|
again = false;
|
|
for (int other = (threadIndex + 1) % JobsUtility.MaxJobThreadCount
|
|
; other != threadIndex
|
|
; other = (other + 1) % JobsUtility.MaxJobThreadCount
|
|
)
|
|
{
|
|
// Attempt to grab a free entry from another thread and switch the other thread's free head
|
|
// atomically.
|
|
do
|
|
{
|
|
idx = data->firstFreeTLS[other * UnsafeParallelHashMapData.IntsPerCacheLine];
|
|
|
|
if (idx < 0)
|
|
{
|
|
break;
|
|
}
|
|
}
|
|
while (Interlocked.CompareExchange(
|
|
ref data->firstFreeTLS[other * UnsafeParallelHashMapData.IntsPerCacheLine]
|
|
, nextPtrs[idx]
|
|
, idx
|
|
) != idx
|
|
);
|
|
|
|
if (idx == -2)
|
|
{
|
|
// If the thread was refilling the cache, then try again.
|
|
again = true;
|
|
}
|
|
else if (idx >= 0)
|
|
{
|
|
// We succeeded in getting an entry from another thread so remove this entry from the
|
|
// linked list.
|
|
nextPtrs[idx] = -1;
|
|
return idx;
|
|
}
|
|
}
|
|
}
|
|
ThrowFull();
|
|
}
|
|
|
|
CheckOutOfCapacity(idx, data->keyCapacity);
|
|
}
|
|
while (Interlocked.CompareExchange(
|
|
ref data->firstFreeTLS[threadIndex * UnsafeParallelHashMapData.IntsPerCacheLine]
|
|
, nextPtrs[idx]
|
|
, idx
|
|
) != idx
|
|
);
|
|
|
|
nextPtrs[idx] = -1;
|
|
return idx;
|
|
}
|
|
|
|
internal static unsafe void FreeEntry(UnsafeParallelHashMapData* data, int idx, int threadIndex)
|
|
{
|
|
int* nextPtrs = (int*)data->next;
|
|
int next = -1;
|
|
|
|
do
|
|
{
|
|
next = data->firstFreeTLS[threadIndex * UnsafeParallelHashMapData.IntsPerCacheLine];
|
|
nextPtrs[idx] = next;
|
|
}
|
|
while (Interlocked.CompareExchange(
|
|
ref data->firstFreeTLS[threadIndex * UnsafeParallelHashMapData.IntsPerCacheLine]
|
|
, idx
|
|
, next
|
|
) != next
|
|
);
|
|
}
|
|
|
|
internal static unsafe bool TryAddAtomic(UnsafeParallelHashMapData* data, TKey key, TValue item, int threadIndex)
|
|
{
|
|
TValue tempItem;
|
|
NativeParallelMultiHashMapIterator<TKey> tempIt;
|
|
if (TryGetFirstValueAtomic(data, key, out tempItem, out tempIt))
|
|
{
|
|
return false;
|
|
}
|
|
|
|
// Allocate an entry from the free list
|
|
int idx = AllocEntry(data, threadIndex);
|
|
|
|
// Write the new value to the entry
|
|
UnsafeUtility.WriteArrayElement(data->keys, idx, key);
|
|
UnsafeUtility.WriteArrayElement(data->values, idx, item);
|
|
|
|
int bucket = key.GetHashCode() & data->bucketCapacityMask;
|
|
// Add the index to the hash-map
|
|
int* buckets = (int*)data->buckets;
|
|
|
|
// Make the bucket's head idx. If the exchange returns something other than -1, then the bucket had
|
|
// a non-null head which means we need to do more checks...
|
|
if (Interlocked.CompareExchange(ref buckets[bucket], idx, -1) != -1)
|
|
{
|
|
int* nextPtrs = (int*)data->next;
|
|
int next = -1;
|
|
|
|
do
|
|
{
|
|
// Link up this entry with the rest of the bucket under the assumption that this key
|
|
// doesn't already exist in the bucket. This assumption could be wrong, which will be
|
|
// checked later.
|
|
next = buckets[bucket];
|
|
nextPtrs[idx] = next;
|
|
|
|
// If the key already exists then we should free the entry we took earlier.
|
|
if (TryGetFirstValueAtomic(data, key, out tempItem, out tempIt))
|
|
{
|
|
// Put back the entry in the free list if someone else added it while trying to add
|
|
FreeEntry(data, idx, threadIndex);
|
|
|
|
return false;
|
|
}
|
|
}
|
|
while (Interlocked.CompareExchange(ref buckets[bucket], idx, next) != next);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
internal static unsafe void AddAtomicMulti(UnsafeParallelHashMapData* data, TKey key, TValue item, int threadIndex)
|
|
{
|
|
// Allocate an entry from the free list
|
|
int idx = AllocEntry(data, threadIndex);
|
|
|
|
// Write the new value to the entry
|
|
UnsafeUtility.WriteArrayElement(data->keys, idx, key);
|
|
UnsafeUtility.WriteArrayElement(data->values, idx, item);
|
|
|
|
int bucket = key.GetHashCode() & data->bucketCapacityMask;
|
|
// Add the index to the hash-map
|
|
int* buckets = (int*)data->buckets;
|
|
|
|
int nextPtr;
|
|
int* nextPtrs = (int*)data->next;
|
|
do
|
|
{
|
|
nextPtr = buckets[bucket];
|
|
nextPtrs[idx] = nextPtr;
|
|
}
|
|
while (Interlocked.CompareExchange(ref buckets[bucket], idx, nextPtr) != nextPtr);
|
|
}
|
|
|
|
internal static unsafe bool TryAdd(UnsafeParallelHashMapData* data, TKey key, TValue item, bool isMultiHashMap, AllocatorManager.AllocatorHandle allocation)
|
|
{
|
|
TValue tempItem;
|
|
NativeParallelMultiHashMapIterator<TKey> tempIt;
|
|
if (!isMultiHashMap && TryGetFirstValueAtomic(data, key, out tempItem, out tempIt))
|
|
{
|
|
return false;
|
|
}
|
|
|
|
// Allocate an entry from the free list
|
|
int idx;
|
|
int* nextPtrs;
|
|
|
|
if (data->allocatedIndexLength >= data->keyCapacity && data->firstFreeTLS[0] < 0)
|
|
{
|
|
for (int tls = 1; tls < JobsUtility.MaxJobThreadCount; ++tls)
|
|
{
|
|
if (data->firstFreeTLS[tls * UnsafeParallelHashMapData.IntsPerCacheLine] >= 0)
|
|
{
|
|
idx = data->firstFreeTLS[tls * UnsafeParallelHashMapData.IntsPerCacheLine];
|
|
nextPtrs = (int*)data->next;
|
|
data->firstFreeTLS[tls * UnsafeParallelHashMapData.IntsPerCacheLine] = nextPtrs[idx];
|
|
nextPtrs[idx] = -1;
|
|
data->firstFreeTLS[0] = idx;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (data->firstFreeTLS[0] < 0)
|
|
{
|
|
int newCap = UnsafeParallelHashMapData.GrowCapacity(data->keyCapacity);
|
|
UnsafeParallelHashMapData.ReallocateHashMap<TKey, TValue>(data, newCap, UnsafeParallelHashMapData.GetBucketSize(newCap), allocation);
|
|
}
|
|
}
|
|
|
|
idx = data->firstFreeTLS[0];
|
|
|
|
if (idx >= 0)
|
|
{
|
|
data->firstFreeTLS[0] = ((int*)data->next)[idx];
|
|
}
|
|
else
|
|
{
|
|
idx = data->allocatedIndexLength++;
|
|
}
|
|
|
|
CheckIndexOutOfBounds(data, idx);
|
|
|
|
// Write the new value to the entry
|
|
UnsafeUtility.WriteArrayElement(data->keys, idx, key);
|
|
UnsafeUtility.WriteArrayElement(data->values, idx, item);
|
|
|
|
int bucket = key.GetHashCode() & data->bucketCapacityMask;
|
|
// Add the index to the hash-map
|
|
int* buckets = (int*)data->buckets;
|
|
nextPtrs = (int*)data->next;
|
|
nextPtrs[idx] = buckets[bucket];
|
|
buckets[bucket] = idx;
|
|
|
|
return true;
|
|
}
|
|
|
|
internal static unsafe int Remove(UnsafeParallelHashMapData* data, TKey key, bool isMultiHashMap)
|
|
{
|
|
if (data->keyCapacity == 0)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
var removed = 0;
|
|
|
|
// First find the slot based on the hash
|
|
var buckets = (int*)data->buckets;
|
|
var nextPtrs = (int*)data->next;
|
|
var bucket = key.GetHashCode() & data->bucketCapacityMask;
|
|
var prevEntry = -1;
|
|
var entryIdx = buckets[bucket];
|
|
|
|
while (entryIdx >= 0 && entryIdx < data->keyCapacity)
|
|
{
|
|
if (UnsafeUtility.ReadArrayElement<TKey>(data->keys, entryIdx).Equals(key))
|
|
{
|
|
++removed;
|
|
|
|
// Found matching element, remove it
|
|
if (prevEntry < 0)
|
|
{
|
|
buckets[bucket] = nextPtrs[entryIdx];
|
|
}
|
|
else
|
|
{
|
|
nextPtrs[prevEntry] = nextPtrs[entryIdx];
|
|
}
|
|
|
|
// And free the index
|
|
int nextIdx = nextPtrs[entryIdx];
|
|
nextPtrs[entryIdx] = data->firstFreeTLS[0];
|
|
data->firstFreeTLS[0] = entryIdx;
|
|
entryIdx = nextIdx;
|
|
|
|
// Can only be one hit in regular hashmaps, so return
|
|
if (!isMultiHashMap)
|
|
{
|
|
break;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
prevEntry = entryIdx;
|
|
entryIdx = nextPtrs[entryIdx];
|
|
}
|
|
}
|
|
|
|
return removed;
|
|
}
|
|
|
|
internal static unsafe void Remove(UnsafeParallelHashMapData* data, NativeParallelMultiHashMapIterator<TKey> it)
|
|
{
|
|
// First find the slot based on the hash
|
|
int* buckets = (int*)data->buckets;
|
|
int* nextPtrs = (int*)data->next;
|
|
int bucket = it.key.GetHashCode() & data->bucketCapacityMask;
|
|
|
|
int entryIdx = buckets[bucket];
|
|
|
|
if (entryIdx == it.EntryIndex)
|
|
{
|
|
buckets[bucket] = nextPtrs[entryIdx];
|
|
}
|
|
else
|
|
{
|
|
while (entryIdx >= 0 && nextPtrs[entryIdx] != it.EntryIndex)
|
|
{
|
|
entryIdx = nextPtrs[entryIdx];
|
|
}
|
|
|
|
if (entryIdx < 0)
|
|
{
|
|
ThrowInvalidIterator();
|
|
}
|
|
|
|
nextPtrs[entryIdx] = nextPtrs[it.EntryIndex];
|
|
}
|
|
|
|
// And free the index
|
|
nextPtrs[it.EntryIndex] = data->firstFreeTLS[0];
|
|
data->firstFreeTLS[0] = it.EntryIndex;
|
|
}
|
|
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int) })]
|
|
internal static unsafe void RemoveKeyValue<TValueEQ>(UnsafeParallelHashMapData* data, TKey key, TValueEQ value)
|
|
where TValueEQ : struct, IEquatable<TValueEQ>
|
|
{
|
|
if (data->keyCapacity == 0)
|
|
{
|
|
return;
|
|
}
|
|
|
|
var buckets = (int*)data->buckets;
|
|
var keyCapacity = (uint)data->keyCapacity;
|
|
var prevNextPtr = buckets + (key.GetHashCode() & data->bucketCapacityMask);
|
|
var entryIdx = *prevNextPtr;
|
|
|
|
if ((uint)entryIdx >= keyCapacity)
|
|
{
|
|
return;
|
|
}
|
|
|
|
var nextPtrs = (int*)data->next;
|
|
var keys = data->keys;
|
|
var values = data->values;
|
|
var firstFreeTLS = data->firstFreeTLS;
|
|
|
|
do
|
|
{
|
|
if (UnsafeUtility.ReadArrayElement<TKey>(keys, entryIdx).Equals(key)
|
|
&& UnsafeUtility.ReadArrayElement<TValueEQ>(values, entryIdx).Equals(value))
|
|
{
|
|
int nextIdx = nextPtrs[entryIdx];
|
|
nextPtrs[entryIdx] = firstFreeTLS[0];
|
|
firstFreeTLS[0] = entryIdx;
|
|
*prevNextPtr = entryIdx = nextIdx;
|
|
}
|
|
else
|
|
{
|
|
prevNextPtr = nextPtrs + entryIdx;
|
|
entryIdx = *prevNextPtr;
|
|
}
|
|
}
|
|
while ((uint)entryIdx < keyCapacity);
|
|
}
|
|
|
|
internal static unsafe bool TryGetFirstValueAtomic(UnsafeParallelHashMapData* data, TKey key, out TValue item, out NativeParallelMultiHashMapIterator<TKey> it)
|
|
{
|
|
it.key = key;
|
|
|
|
if (data->allocatedIndexLength <= 0)
|
|
{
|
|
it.EntryIndex = it.NextEntryIndex = -1;
|
|
item = default;
|
|
return false;
|
|
}
|
|
|
|
// First find the slot based on the hash
|
|
int* buckets = (int*)data->buckets;
|
|
int bucket = key.GetHashCode() & data->bucketCapacityMask;
|
|
it.EntryIndex = it.NextEntryIndex = buckets[bucket];
|
|
return TryGetNextValueAtomic(data, out item, ref it);
|
|
}
|
|
|
|
internal static unsafe bool TryGetNextValueAtomic(UnsafeParallelHashMapData* data, out TValue item, ref NativeParallelMultiHashMapIterator<TKey> it)
|
|
{
|
|
int entryIdx = it.NextEntryIndex;
|
|
it.NextEntryIndex = -1;
|
|
it.EntryIndex = -1;
|
|
item = default;
|
|
if (entryIdx < 0 || entryIdx >= data->keyCapacity)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
int* nextPtrs = (int*)data->next;
|
|
while (!UnsafeUtility.ReadArrayElement<TKey>(data->keys, entryIdx).Equals(it.key))
|
|
{
|
|
entryIdx = nextPtrs[entryIdx];
|
|
if (entryIdx < 0 || entryIdx >= data->keyCapacity)
|
|
{
|
|
return false;
|
|
}
|
|
}
|
|
|
|
it.NextEntryIndex = nextPtrs[entryIdx];
|
|
it.EntryIndex = entryIdx;
|
|
|
|
// Read the value
|
|
item = UnsafeUtility.ReadArrayElement<TValue>(data->values, entryIdx);
|
|
|
|
return true;
|
|
}
|
|
|
|
internal static unsafe bool SetValue(UnsafeParallelHashMapData* data, ref NativeParallelMultiHashMapIterator<TKey> it, ref TValue item)
|
|
{
|
|
int entryIdx = it.EntryIndex;
|
|
if (entryIdx < 0 || entryIdx >= data->keyCapacity)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
UnsafeUtility.WriteArrayElement(data->values, entryIdx, item);
|
|
return true;
|
|
}
|
|
|
|
[Conditional("ENABLE_UNITY_COLLECTIONS_CHECKS")]
|
|
static void CheckOutOfCapacity(int idx, int keyCapacity)
|
|
{
|
|
if (idx >= keyCapacity)
|
|
{
|
|
throw new InvalidOperationException(string.Format("nextPtr idx {0} beyond capacity {1}", idx, keyCapacity));
|
|
}
|
|
}
|
|
|
|
[Conditional("ENABLE_UNITY_COLLECTIONS_CHECKS")]
|
|
static unsafe void CheckIndexOutOfBounds(UnsafeParallelHashMapData* data, int idx)
|
|
{
|
|
if (idx < 0 || idx >= data->keyCapacity)
|
|
throw new InvalidOperationException("Internal HashMap error");
|
|
}
|
|
|
|
[Conditional("ENABLE_UNITY_COLLECTIONS_CHECKS")]
|
|
static void ThrowFull()
|
|
{
|
|
throw new InvalidOperationException("HashMap is full");
|
|
}
|
|
|
|
[Conditional("ENABLE_UNITY_COLLECTIONS_CHECKS")]
|
|
static void ThrowInvalidIterator()
|
|
{
|
|
throw new InvalidOperationException("Invalid iterator passed to HashMap remove");
|
|
}
|
|
}
|
|
|
|
/// <summary>
|
|
/// A key-value pair.
|
|
/// </summary>
|
|
/// <remarks>Used for enumerators.</remarks>
|
|
/// <typeparam name="TKey">The type of the keys.</typeparam>
|
|
/// <typeparam name="TValue">The type of the values.</typeparam>
|
|
[DebuggerDisplay("Key = {Key}, Value = {Value}")]
|
|
[BurstCompatible(GenericTypeArguments = new[] {typeof(int), typeof(int)})]
|
|
public unsafe struct KeyValue<TKey, TValue>
|
|
where TKey : struct, IEquatable<TKey>
|
|
where TValue : struct
|
|
{
|
|
internal UnsafeParallelHashMapData* m_Buffer;
|
|
internal int m_Index;
|
|
internal int m_Next;
|
|
|
|
/// <summary>
|
|
/// An invalid KeyValue.
|
|
/// </summary>
|
|
/// <value>In a hash map enumerator's initial state, its <see cref="UnsafeHashMap{TKey,TValue}.Enumerator.Current"/> value is Null.</value>
|
|
public static KeyValue<TKey, TValue> Null => new KeyValue<TKey, TValue>{m_Index = -1};
|
|
|
|
/// <summary>
|
|
/// The key.
|
|
/// </summary>
|
|
/// <value>The key. If this KeyValue is Null, returns the default of TKey.</value>
|
|
public TKey Key
|
|
{
|
|
get
|
|
{
|
|
if (m_Index != -1)
|
|
{
|
|
return UnsafeUtility.ReadArrayElement<TKey>(m_Buffer->keys, m_Index);
|
|
}
|
|
|
|
return default;
|
|
}
|
|
}
|
|
|
|
/// <summary>
|
|
/// Value of key/value pair.
|
|
/// </summary>
|
|
public ref TValue Value
|
|
{
|
|
get
|
|
{
|
|
#if ENABLE_UNITY_COLLECTIONS_CHECKS
|
|
if (m_Index == -1)
|
|
throw new ArgumentException("must be valid");
|
|
#endif
|
|
|
|
return ref UnsafeUtility.AsRef<TValue>(m_Buffer->values + UnsafeUtility.SizeOf<TValue>() * m_Index);
|
|
}
|
|
}
|
|
|
|
/// <summary>
|
|
/// Gets the key and the value.
|
|
/// </summary>
|
|
/// <param name="key">Outputs the key. If this KeyValue is Null, outputs the default of TKey.</param>
|
|
/// <param name="value">Outputs the value. If this KeyValue is Null, outputs the default of TValue.</param>
|
|
/// <returns>True if the key-value pair is valid.</returns>
|
|
public bool GetKeyValue(out TKey key, out TValue value)
|
|
{
|
|
if (m_Index != -1)
|
|
{
|
|
key = UnsafeUtility.ReadArrayElement<TKey>(m_Buffer->keys, m_Index);
|
|
value = UnsafeUtility.ReadArrayElement<TValue>(m_Buffer->values, m_Index);
|
|
return true;
|
|
}
|
|
|
|
key = default;
|
|
value = default;
|
|
return false;
|
|
}
|
|
}
|
|
|
|
internal unsafe struct UnsafeParallelHashMapDataEnumerator
|
|
{
|
|
[NativeDisableUnsafePtrRestriction]
|
|
internal UnsafeParallelHashMapData* m_Buffer;
|
|
internal int m_Index;
|
|
internal int m_BucketIndex;
|
|
internal int m_NextIndex;
|
|
|
|
internal unsafe UnsafeParallelHashMapDataEnumerator(UnsafeParallelHashMapData* data)
|
|
{
|
|
m_Buffer = data;
|
|
m_Index = -1;
|
|
m_BucketIndex = 0;
|
|
m_NextIndex = -1;
|
|
}
|
|
|
|
internal bool MoveNext()
|
|
{
|
|
return UnsafeParallelHashMapData.MoveNext(m_Buffer, ref m_BucketIndex, ref m_NextIndex, out m_Index);
|
|
}
|
|
|
|
internal void Reset()
|
|
{
|
|
m_Index = -1;
|
|
m_BucketIndex = 0;
|
|
m_NextIndex = -1;
|
|
}
|
|
|
|
internal KeyValue<TKey, TValue> GetCurrent<TKey, TValue>()
|
|
where TKey : struct, IEquatable<TKey>
|
|
where TValue : struct
|
|
{
|
|
return new KeyValue<TKey, TValue> { m_Buffer = m_Buffer, m_Index = m_Index };
|
|
}
|
|
|
|
internal TKey GetCurrentKey<TKey>()
|
|
where TKey : struct, IEquatable<TKey>
|
|
{
|
|
if (m_Index != -1)
|
|
{
|
|
return UnsafeUtility.ReadArrayElement<TKey>(m_Buffer->keys, m_Index);
|
|
}
|
|
|
|
return default;
|
|
}
|
|
}
|
|
|
|
/// <summary>
|
|
/// An unordered, expandable associative array.
|
|
/// </summary>
|
|
/// <typeparam name="TKey">The type of the keys.</typeparam>
|
|
/// <typeparam name="TValue">The type of the values.</typeparam>
|
|
[StructLayout(LayoutKind.Sequential)]
|
|
[DebuggerDisplay("Count = {Count()}, Capacity = {Capacity}, IsCreated = {IsCreated}, IsEmpty = {IsEmpty}")]
|
|
[DebuggerTypeProxy(typeof(UnsafeParallelHashMapDebuggerTypeProxy<,>))]
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int), typeof(int) })]
|
|
public unsafe struct UnsafeParallelHashMap<TKey, TValue>
|
|
: INativeDisposable
|
|
, IEnumerable<KeyValue<TKey, TValue>> // Used by collection initializers.
|
|
where TKey : struct, IEquatable<TKey>
|
|
where TValue : struct
|
|
{
|
|
[NativeDisableUnsafePtrRestriction]
|
|
internal UnsafeParallelHashMapData* m_Buffer;
|
|
internal AllocatorManager.AllocatorHandle m_AllocatorLabel;
|
|
|
|
/// <summary>
|
|
/// Initializes and returns an instance of UnsafeHashMap.
|
|
/// </summary>
|
|
/// <param name="capacity">The number of key-value pairs that should fit in the initial allocation.</param>
|
|
/// <param name="allocator">The allocator to use.</param>
|
|
public UnsafeParallelHashMap(int capacity, AllocatorManager.AllocatorHandle allocator)
|
|
{
|
|
CollectionHelper.CheckIsUnmanaged<TKey>();
|
|
CollectionHelper.CheckIsUnmanaged<TValue>();
|
|
|
|
m_AllocatorLabel = allocator;
|
|
// Bucket size if bigger to reduce collisions
|
|
UnsafeParallelHashMapData.AllocateHashMap<TKey, TValue>(capacity, capacity * 2, allocator, out m_Buffer);
|
|
|
|
Clear();
|
|
}
|
|
|
|
/// <summary>
|
|
/// Whether this hash map is empty.
|
|
/// </summary>
|
|
/// <value>True if this hash map is empty or the hash map has not been constructed.</value>
|
|
public bool IsEmpty => !IsCreated || UnsafeParallelHashMapData.IsEmpty(m_Buffer);
|
|
|
|
/// <summary>
|
|
/// The current number of key-value pairs in this hash map.
|
|
/// </summary>
|
|
/// <returns>The current number of key-value pairs in this hash map.</returns>
|
|
public int Count() => UnsafeParallelHashMapData.GetCount(m_Buffer);
|
|
|
|
/// <summary>
|
|
/// The number of key-value pairs that fit in the current allocation.
|
|
/// </summary>
|
|
/// <value>The number of key-value pairs that fit in the current allocation.</value>
|
|
/// <param name="value">A new capacity. Must be larger than the current capacity.</param>
|
|
/// <exception cref="Exception">Thrown if `value` is less than the current capacity.</exception>
|
|
public int Capacity
|
|
{
|
|
get
|
|
{
|
|
UnsafeParallelHashMapData* data = m_Buffer;
|
|
return data->keyCapacity;
|
|
}
|
|
|
|
set
|
|
{
|
|
UnsafeParallelHashMapData* data = m_Buffer;
|
|
UnsafeParallelHashMapData.ReallocateHashMap<TKey, TValue>(data, value, UnsafeParallelHashMapData.GetBucketSize(value), m_AllocatorLabel);
|
|
}
|
|
}
|
|
|
|
/// <summary>
|
|
/// Removes all key-value pairs.
|
|
/// </summary>
|
|
/// <remarks>Does not change the capacity.</remarks>
|
|
public void Clear()
|
|
{
|
|
UnsafeParallelHashMapBase<TKey, TValue>.Clear(m_Buffer);
|
|
}
|
|
|
|
/// <summary>
|
|
/// Adds a new key-value pair.
|
|
/// </summary>
|
|
/// <remarks>If the key is already present, this method returns false without modifying the hash map.</remarks>
|
|
/// <param name="key">The key to add.</param>
|
|
/// <param name="item">The value to add.</param>
|
|
/// <returns>True if the key-value pair was added.</returns>
|
|
public bool TryAdd(TKey key, TValue item)
|
|
{
|
|
return UnsafeParallelHashMapBase<TKey, TValue>.TryAdd(m_Buffer, key, item, false, m_AllocatorLabel);
|
|
}
|
|
|
|
/// <summary>
|
|
/// Adds a new key-value pair.
|
|
/// </summary>
|
|
/// <remarks>If the key is already present, this method throws without modifying the hash map.</remarks>
|
|
/// <param name="key">The key to add.</param>
|
|
/// <param name="item">The value to add.</param>
|
|
/// <exception cref="ArgumentException">Thrown if the key was already present.</exception>
|
|
public void Add(TKey key, TValue item)
|
|
{
|
|
TryAdd(key, item);
|
|
}
|
|
|
|
/// <summary>
|
|
/// Removes a key-value pair.
|
|
/// </summary>
|
|
/// <param name="key">The key to remove.</param>
|
|
/// <returns>True if a key-value pair was removed.</returns>
|
|
public bool Remove(TKey key)
|
|
{
|
|
return UnsafeParallelHashMapBase<TKey, TValue>.Remove(m_Buffer, key, false) != 0;
|
|
}
|
|
|
|
/// <summary>
|
|
/// Returns the value associated with a key.
|
|
/// </summary>
|
|
/// <param name="key">The key to look up.</param>
|
|
/// <param name="item">Outputs the value associated with the key. Outputs default if the key was not present.</param>
|
|
/// <returns>True if the key was present.</returns>
|
|
public bool TryGetValue(TKey key, out TValue item)
|
|
{
|
|
NativeParallelMultiHashMapIterator<TKey> tempIt;
|
|
return UnsafeParallelHashMapBase<TKey, TValue>.TryGetFirstValueAtomic(m_Buffer, key, out item, out tempIt);
|
|
}
|
|
|
|
/// <summary>
|
|
/// Returns true if a given key is present in this hash map.
|
|
/// </summary>
|
|
/// <param name="key">The key to look up.</param>
|
|
/// <returns>True if the key was present.</returns>
|
|
public bool ContainsKey(TKey key)
|
|
{
|
|
return UnsafeParallelHashMapBase<TKey, TValue>.TryGetFirstValueAtomic(m_Buffer, key, out var tempValue, out var tempIt);
|
|
}
|
|
|
|
/// <summary>
|
|
/// Gets and sets values by key.
|
|
/// </summary>
|
|
/// <remarks>Getting a key that is not present will throw. Setting a key that is not already present will add the key.</remarks>
|
|
/// <param name="key">The key to look up.</param>
|
|
/// <value>The value associated with the key.</value>
|
|
/// <exception cref="ArgumentException">For getting, thrown if the key was not present.</exception>
|
|
public TValue this[TKey key]
|
|
{
|
|
get
|
|
{
|
|
TValue res;
|
|
TryGetValue(key, out res);
|
|
return res;
|
|
}
|
|
|
|
set
|
|
{
|
|
if (UnsafeParallelHashMapBase<TKey, TValue>.TryGetFirstValueAtomic(m_Buffer, key, out var item, out var iterator))
|
|
{
|
|
UnsafeParallelHashMapBase<TKey, TValue>.SetValue(m_Buffer, ref iterator, ref value);
|
|
}
|
|
else
|
|
{
|
|
UnsafeParallelHashMapBase<TKey, TValue>.TryAdd(m_Buffer, key, value, false, m_AllocatorLabel);
|
|
}
|
|
}
|
|
}
|
|
|
|
/// <summary>
|
|
/// Whether this hash map has been allocated (and not yet deallocated).
|
|
/// </summary>
|
|
/// <value>True if this hash map has been allocated (and not yet deallocated).</value>
|
|
public bool IsCreated => m_Buffer != null;
|
|
|
|
/// <summary>
|
|
/// Releases all resources (memory).
|
|
/// </summary>
|
|
public void Dispose()
|
|
{
|
|
UnsafeParallelHashMapData.DeallocateHashMap(m_Buffer, m_AllocatorLabel);
|
|
m_Buffer = null;
|
|
}
|
|
|
|
/// <summary>
|
|
/// Creates and schedules a job that will dispose this hash map.
|
|
/// </summary>
|
|
/// <param name="inputDeps">A job handle. The newly scheduled job will depend upon this handle.</param>
|
|
/// <returns>The handle of a new job that will dispose this hash map.</returns>
|
|
[NotBurstCompatible /* This is not burst compatible because of IJob's use of a static IntPtr. Should switch to IJobBurstSchedulable in the future */]
|
|
public JobHandle Dispose(JobHandle inputDeps)
|
|
{
|
|
var jobHandle = new UnsafeParallelHashMapDisposeJob { Data = m_Buffer, Allocator = m_AllocatorLabel }.Schedule(inputDeps);
|
|
m_Buffer = null;
|
|
return jobHandle;
|
|
}
|
|
|
|
/// <summary>
|
|
/// Returns an array with a copy of all this hash map's keys (in no particular order).
|
|
/// </summary>
|
|
/// <param name="allocator">The allocator to use.</param>
|
|
/// <returns>An array with a copy of all this hash map's keys (in no particular order).</returns>
|
|
public NativeArray<TKey> GetKeyArray(AllocatorManager.AllocatorHandle allocator)
|
|
{
|
|
var result = CollectionHelper.CreateNativeArray<TKey>(UnsafeParallelHashMapData.GetCount(m_Buffer), allocator, NativeArrayOptions.UninitializedMemory);
|
|
UnsafeParallelHashMapData.GetKeyArray(m_Buffer, result);
|
|
return result;
|
|
}
|
|
|
|
/// <summary>
|
|
/// Returns an array with a copy of all this hash map's values (in no particular order).
|
|
/// </summary>
|
|
/// <param name="allocator">The allocator to use.</param>
|
|
/// <returns>An array with a copy of all this hash map's values (in no particular order).</returns>
|
|
public NativeArray<TValue> GetValueArray(AllocatorManager.AllocatorHandle allocator)
|
|
{
|
|
var result = CollectionHelper.CreateNativeArray<TValue>(UnsafeParallelHashMapData.GetCount(m_Buffer), allocator, NativeArrayOptions.UninitializedMemory);
|
|
UnsafeParallelHashMapData.GetValueArray(m_Buffer, result);
|
|
return result;
|
|
}
|
|
|
|
/// <summary>
|
|
/// Returns a NativeKeyValueArrays with a copy of all this hash map's keys and values.
|
|
/// </summary>
|
|
/// <remarks>The key-value pairs are copied in no particular order. For all `i`, `Values[i]` will be the value associated with `Keys[i]`.</remarks>
|
|
/// <param name="allocator">The allocator to use.</param>
|
|
/// <returns>A NativeKeyValueArrays with a copy of all this hash map's keys and values.</returns>
|
|
public NativeKeyValueArrays<TKey, TValue> GetKeyValueArrays(AllocatorManager.AllocatorHandle allocator)
|
|
{
|
|
var result = new NativeKeyValueArrays<TKey, TValue>(UnsafeParallelHashMapData.GetCount(m_Buffer), allocator, NativeArrayOptions.UninitializedMemory);
|
|
UnsafeParallelHashMapData.GetKeyValueArrays(m_Buffer, result);
|
|
return result;
|
|
}
|
|
|
|
/// <summary>
|
|
/// Returns a parallel writer for this hash map.
|
|
/// </summary>
|
|
/// <returns>A parallel writer for this hash map.</returns>
|
|
public ParallelWriter AsParallelWriter()
|
|
{
|
|
ParallelWriter writer;
|
|
writer.m_ThreadIndex = 0;
|
|
writer.m_Buffer = m_Buffer;
|
|
return writer;
|
|
}
|
|
|
|
/// <summary>
|
|
/// A parallel writer for a UnsafeParallelHashMap.
|
|
/// </summary>
|
|
/// <remarks>
|
|
/// Use <see cref="AsParallelWriter"/> to create a parallel writer for a UnsafeParallelHashMap.
|
|
/// </remarks>
|
|
[NativeContainerIsAtomicWriteOnly]
|
|
[BurstCompatible(GenericTypeArguments = new [] { typeof(int), typeof(int) })]
|
|
public unsafe struct ParallelWriter
|
|
{
|
|
[NativeDisableUnsafePtrRestriction]
|
|
internal UnsafeParallelHashMapData* m_Buffer;
|
|
|
|
[NativeSetThreadIndex]
|
|
internal int m_ThreadIndex;
|
|
|
|
/// <summary>
|
|
/// The number of key-value pairs that fit in the current allocation.
|
|
/// </summary>
|
|
/// <value>The number of key-value pairs that fit in the current allocation.</value>
|
|
public int Capacity
|
|
{
|
|
get
|
|
{
|
|
UnsafeParallelHashMapData* data = m_Buffer;
|
|
return data->keyCapacity;
|
|
}
|
|
}
|
|
|
|
/// <summary>
|
|
/// Adds a new key-value pair.
|
|
/// </summary>
|
|
/// <remarks>If the key is already present, this method returns false without modifying the hash map.</remarks>
|
|
/// <param name="key">The key to add.</param>
|
|
/// <param name="item">The value to add.</param>
|
|
/// <returns>True if the key-value pair was added.</returns>
|
|
public bool TryAdd(TKey key, TValue item)
|
|
{
|
|
Assert.IsTrue(m_ThreadIndex >= 0);
|
|
return UnsafeParallelHashMapBase<TKey, TValue>.TryAddAtomic(m_Buffer, key, item, m_ThreadIndex);
|
|
}
|
|
}
|
|
|
|
/// <summary>
|
|
/// Returns an enumerator over the key-value pairs of this hash map.
|
|
/// </summary>
|
|
/// <returns>An enumerator over the key-value pairs of this hash map.</returns>
|
|
public Enumerator GetEnumerator()
|
|
{
|
|
return new Enumerator { m_Enumerator = new UnsafeParallelHashMapDataEnumerator(m_Buffer) };
|
|
}
|
|
|
|
/// <summary>
|
|
/// This method is not implemented. Use <see cref="GetEnumerator"/> instead.
|
|
/// </summary>
|
|
/// <returns>Throws NotImplementedException.</returns>
|
|
/// <exception cref="NotImplementedException">Method is not implemented.</exception>
|
|
IEnumerator<KeyValue<TKey, TValue>> IEnumerable<KeyValue<TKey, TValue>>.GetEnumerator()
|
|
{
|
|
throw new NotImplementedException();
|
|
}
|
|
|
|
/// <summary>
|
|
/// This method is not implemented. Use <see cref="GetEnumerator"/> instead.
|
|
/// </summary>
|
|
/// <returns>Throws NotImplementedException.</returns>
|
|
/// <exception cref="NotImplementedException">Method is not implemented.</exception>
|
|
IEnumerator IEnumerable.GetEnumerator()
|
|
{
|
|
throw new NotImplementedException();
|
|
}
|
|
|
|
/// <summary>
|
|
/// An enumerator over the key-value pairs of a hash map.
|
|
/// </summary>
|
|
/// <remarks>
|
|
/// In an enumerator's initial state, <see cref="Current"/> is not valid to read.
|
|
/// From this state, the first <see cref="MoveNext"/> call advances the enumerator to the first key-value pair.
|
|
/// </remarks>
|
|
public struct Enumerator : IEnumerator<KeyValue<TKey, TValue>>
|
|
{
|
|
internal UnsafeParallelHashMapDataEnumerator m_Enumerator;
|
|
|
|
/// <summary>
|
|
/// Does nothing.
|
|
/// </summary>
|
|
public void Dispose() { }
|
|
|
|
/// <summary>
|
|
/// Advances the enumerator to the next key-value pair.
|
|
/// </summary>
|
|
/// <returns>True if <see cref="Current"/> is valid to read after the call.</returns>
|
|
public bool MoveNext() => m_Enumerator.MoveNext();
|
|
|
|
/// <summary>
|
|
/// Resets the enumerator to its initial state.
|
|
/// </summary>
|
|
public void Reset() => m_Enumerator.Reset();
|
|
|
|
/// <summary>
|
|
/// The current key-value pair.
|
|
/// </summary>
|
|
/// <value>The current key-value pair.</value>
|
|
public KeyValue<TKey, TValue> Current => m_Enumerator.GetCurrent<TKey, TValue>();
|
|
|
|
object IEnumerator.Current => Current;
|
|
}
|
|
}
|
|
|
|
[BurstCompile]
|
|
internal unsafe struct UnsafeParallelHashMapDisposeJob : IJob
|
|
{
|
|
[NativeDisableUnsafePtrRestriction]
|
|
public UnsafeParallelHashMapData* Data;
|
|
public AllocatorManager.AllocatorHandle Allocator;
|
|
|
|
public void Execute()
|
|
{
|
|
UnsafeParallelHashMapData.DeallocateHashMap(Data, Allocator);
|
|
}
|
|
}
|
|
|
|
sealed internal class UnsafeParallelHashMapDebuggerTypeProxy<TKey, TValue>
|
|
where TKey : struct, IEquatable<TKey>
|
|
where TValue : struct
|
|
{
|
|
#if !NET_DOTS
|
|
UnsafeParallelHashMap<TKey, TValue> m_Target;
|
|
|
|
public UnsafeParallelHashMapDebuggerTypeProxy(UnsafeParallelHashMap<TKey, TValue> target)
|
|
{
|
|
m_Target = target;
|
|
}
|
|
|
|
public List<Pair<TKey, TValue>> Items
|
|
{
|
|
get
|
|
{
|
|
var result = new List<Pair<TKey, TValue>>();
|
|
using (var kva = m_Target.GetKeyValueArrays(Allocator.Temp))
|
|
{
|
|
for (var i = 0; i < kva.Length; ++i)
|
|
{
|
|
result.Add(new Pair<TKey, TValue>(kva.Keys[i], kva.Values[i]));
|
|
}
|
|
}
|
|
return result;
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/// <summary>
|
|
/// For internal use only.
|
|
/// </summary>
|
|
[StructLayout(LayoutKind.Sequential)]
|
|
public unsafe struct UntypedUnsafeParallelHashMap
|
|
{
|
|
#pragma warning disable 169
|
|
[NativeDisableUnsafePtrRestriction]
|
|
UnsafeParallelHashMapData* m_Buffer;
|
|
AllocatorManager.AllocatorHandle m_AllocatorLabel;
|
|
#pragma warning restore 169
|
|
}
|
|
}
|