// Licensed to the .NET Foundation under one or more agreements. // The .NET Foundation licenses this file to you under the MIT license. // See the LICENSE file in the project root for more information. using System; namespace HeuristicLab.Problems.ProgramSynthesis.Push.Data.Pool { using System.Collections.Generic; using System.Threading; /// /// Generic implementation of object pooling pattern with predefined pool size limit. The main /// purpose is that limited number of frequently used objects can be kept in the pool for /// further recycling. /// /// Notes: /// 1) it is not the goal to keep all returned objects. Pool is not meant for storage. If there /// is no space in the pool, extra returned objects will be dropped. /// /// 2) it is implied that if object was obtained from a pool, the caller will return it back in /// a relatively short time. Keeping checked out objects for long durations is ok, but /// reduces usefulness of pooling. Just new up your own. /// /// Not returning objects to the pool in not detrimental to the pool's work, but is a bad practice. /// Rationale: /// If there is no intent for reusing the object, do not use pool - just use "new". /// public sealed class ObjectPool where T : class { private struct Element { internal T Value; } // storage for the pool objects. private readonly Element[] _items; // factory is stored for the lifetime of the pool. We will call this only when pool needs to // expand. compared to "new T()", Func gives more flexibility to implementers and faster // than "new T()". private readonly Func _factory; public ObjectPool(Func factory) : this(factory, Environment.ProcessorCount * 2) { } public ObjectPool(Func factory, int size) { _factory = factory; _items = new Element[size]; } private T CreateInstance() { var inst = _factory(); return inst; } /// /// Produces an instance. /// /// /// Search strategy is a simple linear probing which is chosen for it cache-friendliness. /// Note that Free will try to store recycled objects close to the start thus statistically /// reducing how far we will typically search. /// public T Allocate() { var items = _items; T inst; for (int i = 0; i < items.Length; i++) { // Note that the read is optimistically not synchronized. That is intentional. // We will interlock only when we have a candidate. in a worst case we may miss some // recently returned objects. Not a big deal. inst = items[i].Value; if (inst != null) { if (inst == Interlocked.CompareExchange(ref items[i].Value, null, inst)) { goto gotInstance; } } } inst = CreateInstance(); gotInstance: return inst; } /// /// Returns objects to the pool. /// /// /// Search strategy is a simple linear probing which is chosen for it cache-friendliness. /// Note that Free will try to store recycled objects close to the start thus statistically /// reducing how far we will typically search in Allocate. /// public void Free(T obj) { var items = _items; for (int i = 0; i < items.Length; i++) { if (items[i].Value == null) { // Intentionally not using interlocked here. // In a worst case scenario two objects may be stored into same slot. // It is very unlikely to happen and will only mean that one of the objects will get collected. items[i].Value = obj; break; } } } public void Clear() { for (int i = 0; i < _items.Length; i++) { _items[i] = default(Element); } } public IEnumerable Items { get { var items = _items; for (var i = 0; i < items.Length; i++) { if (items[i].Value != null) { yield return items[i].Value; } } } } } }