MemoryCache does not obey memory limits in configuration

前端 未结 7 1090
暖寄归人
暖寄归人 2020-11-28 18:01

I’m working with the .NET 4.0 MemoryCache class in an application and trying to limit the maximum cache size, but in my tests it does not appear that the cache is actually o

相关标签:
7条回答
  • 2020-11-28 18:29

    If you use the following modified class and monitor the memory via Task Manager does in fact get trimmed:

    internal class Cache
    {
        private Object Statlock = new object();
        private int ItemCount;
        private long size;
        private MemoryCache MemCache;
        private CacheItemPolicy CIPOL = new CacheItemPolicy();
    
        public Cache(double CacheSize)
        {
            NameValueCollection CacheSettings = new NameValueCollection(3);
            CacheSettings.Add("cacheMemoryLimitMegabytes", Convert.ToString(CacheSize));
            CacheSettings.Add("pollingInterval", Convert.ToString("00:00:01"));
            MemCache = new MemoryCache("TestCache", CacheSettings);
        }
    
        public void AddItem(string Name, string Value)
        {
            CacheItem CI = new CacheItem(Name, Value);
            MemCache.Add(CI, CIPOL);
    
            Console.WriteLine(MemCache.GetCount());
        }
    }
    
    0 讨论(0)
  • 2020-11-28 18:31

    I have done some testing with the example of @Canacourse and the modification of @woany and I think there are some critical calls that block the cleaning of the memory cache.

    public void CacheItemRemoved(CacheEntryRemovedArguments Args)
    {
        // this WriteLine() will block the thread of
        // the MemoryCache long enough to slow it down,
        // and it will never catch up the amount of memory
        // beyond the limit
        Console.WriteLine("...");
    
        // ...
    
        // this ReadKey() will block the thread of 
        // the MemoryCache completely, till you press any key
        Console.ReadKey();
    }
    

    But why does the modification of @woany seems to keep the memory at the same level? Firstly, the RemovedCallback is not set and there is no console output or waiting for input that could block the thread of the memory cache.

    Secondly...

    public void AddItem(string Name, string Value)
    {
        // ...
    
        // this WriteLine will block the main thread long enough,
        // so that the thread of the MemoryCache can do its work more frequently
        Console.WriteLine("...");
    }
    

    A Thread.Sleep(1) every ~1000th AddItem() would have the same effect.

    Well, it's not a very deep investigation of the problem, but it looks as if the thread of the MemoryCache does not get enough CPU time for cleaning, while many new elements are added.

    0 讨论(0)
  • 2020-11-28 18:32

    Wow, so I just spent entirely too much time digging around in the CLR with reflector, but I think I finally have a good handle on what's going on here.

    The settings are being read in correctly, but there seems to be a deep-seated problem in the CLR itself that looks like it will render the memory limit setting essentially useless.

    The following code is reflected out of the System.Runtime.Caching DLL, for the CacheMemoryMonitor class (there is a similar class that monitors physical memory and deals with the other setting, but this is the more important one):

    protected override int GetCurrentPressure()
    {
      int num = GC.CollectionCount(2);
      SRef ref2 = this._sizedRef;
      if ((num != this._gen2Count) && (ref2 != null))
      {
        this._gen2Count = num;
        this._idx ^= 1;
        this._cacheSizeSampleTimes[this._idx] = DateTime.UtcNow;
        this._cacheSizeSamples[this._idx] = ref2.ApproximateSize;
        IMemoryCacheManager manager = s_memoryCacheManager;
        if (manager != null)
        {
          manager.UpdateCacheSize(this._cacheSizeSamples[this._idx], this._memoryCache);
        }
      }
      if (this._memoryLimit <= 0L)
      {
        return 0;
      }
      long num2 = this._cacheSizeSamples[this._idx];
      if (num2 > this._memoryLimit)
      {
        num2 = this._memoryLimit;
      }
      return (int) ((num2 * 100L) / this._memoryLimit);
    }
    

    The first thing you might notice is that it doesn't even try to look at the size of the cache until after a Gen2 garbage collection, instead just falling back on the existing stored size value in cacheSizeSamples. So you won't ever be able to hit the target right on, but if the rest worked we would at least get a size measurement before we got in real trouble.

    So assuming a Gen2 GC has occurred, we run into problem 2, which is that ref2.ApproximateSize does a horrible job of actually approximating the size of the cache. Slogging through CLR junk I found that this is a System.SizedReference, and this is what it's doing to get the value (IntPtr is a handle to the MemoryCache object itself):

    [SecurityCritical]
    [MethodImpl(MethodImplOptions.InternalCall)]
    private static extern long GetApproximateSizeOfSizedRef(IntPtr h);
    

    I'm assuming that extern declaration means that it goes diving into unmanaged windows land at this point, and I have no idea how to start finding out what it does there. From what I've observed though it does a horrible job of trying to approximate the size of the overall thing.

    The third noticeable thing there is the call to manager.UpdateCacheSize which sounds like it should do something. Unfortunately in any normal sample of how this should work s_memoryCacheManager will always be null. The field is set from the public static member ObjectCache.Host. This is exposed for the user to mess with if he so chooses, and I was actually able to make this thing sort of work like it's supposed to by slopping together my own IMemoryCacheManager implementation, setting it to ObjectCache.Host, and then running the sample. At that point though, it seems like you might as well just make your own cache implementation and not even bother with all this stuff, especially since I have no idea if setting your own class to ObjectCache.Host (static, so it affects every one of these that might be out there in process) to measure the cache could mess up other things.

    I have to believe that at least part of this (if not a couple parts) is just a straight up bug. It'd be nice to hear from someone at MS what the deal was with this thing.

    TLDR version of this giant answer: Assume that CacheMemoryLimitMegabytes is completely busted at this point in time. You can set it to 10 MB, and then proceed to fill up the cache to ~2GB and blow an out of memory exception with no tripping of item removal.

    0 讨论(0)
  • 2020-11-28 18:32

    I know this answer is crazy late, but better late than never. I wanted to let you know that I wrote a version of MemoryCache that resolves the Gen 2 Collection issues automatically for you. It therefore trims whenever the polling interval indicates memory pressure. If you're experiencing this issue, give it a go!

    http://www.nuget.org/packages/SharpMemoryCache

    You can also find it on GitHub if you're curious about how I solved it. The code is somewhat simple.

    https://github.com/haneytron/sharpmemorycache

    0 讨论(0)
  • 2020-11-28 18:33

    I (thankfully) stumbled across this useful post yesterday when first attempting to use the MemoryCache. I thought it would be a simple case of setting values and using the classes but I encountered similar issues outlined above. To try and see what was going on I extracted the source using ILSpy and then set up a test and stepped through the code. My test code was very similar to the code above so I won't post it. From my tests I noticed that the measurement of the cache size was never particularly accurate (as mentioned above) and given the current implementation would never work reliably. However the physical measurement was fine and if the physical memory was measured at every poll then it seemed to me like the code would work reliably. So, I removed the gen 2 garbage collection check within MemoryCacheStatistics; under normal conditions no memory measurements will be taken unless there has been another gen 2 garbage collection since the last measurement.

    In a test scenario this obviously makes a big difference as the cache is being hit constantly so objects never have the chance to get to gen 2. I think we are going to use the modified build of this dll on our project and use the official MS build when .net 4.5 comes out (which according to the connect article mentioned above should have the fix in it). Logically I can see why the gen 2 check has been put in place but in practise I'm not sure if it makes much sense. If the memory reaches 90% (or whatever limit it has been set to) then it should not matter if a gen 2 collection has occured or not, items should be evicted regardless.

    I left my test code running for about 15 minutes with a the physicalMemoryLimitPercentage set to 65%. I saw the memory usage remain between 65-68% during the test and saw things getting evicted properly. In my test I set the pollingInterval to 5 seconds, physicalMemoryLimitPercentage to 65 and physicalMemoryLimitPercentage to 0 to default this.

    Following the above advice; an implementation of IMemoryCacheManager could be made to evict things from the cache. It would however suffer from the gen 2 check issue mentioned. Although, depending on the scenario, this may not be a problem in production code and may work sufficiently for people.

    0 讨论(0)
  • 2020-11-28 18:37

    It turned out it is not a bug , all what you need to do is setting the pooling time span to enforce the limits , it seem if you leave the pooling not set, it will never trigger.I just tested it and no need to wrappers or any extra code :

     private static readonly NameValueCollection Collection = new NameValueCollection
            {
                {"CacheMemoryLimitMegabytes", "20"},
               {"PollingInterval", TimeSpan.FromMilliseconds(60000).ToString()}, // this will check the limits each 60 seconds
    
            };
    

    Set the value of "PollingInterval" based on how fast the cache is growing , if it grow too fast increase the frequency of polling checks otherwise keep the checks not very frequent to not cause overhead.

    0 讨论(0)
提交回复
热议问题