2 Replies Latest reply on Jun 13, 2013 5:56 AM by James.zhang

    Consolidation performance

      Dear Experts,

      I am aware of MaxNumDataRecordsInRAM and MaxDataCacheSizeinMB settings.
      Having bumped up those settings to upper values (40M and 6000 currently), we can see during busy times HFM application (Hsvdatasource.exe) can take 30 GB of 32 GB RAM in the application server. Performance tuning guide suggests we could even increase more like 60M and 9000 if we have 32 GB of RAM given that only memory intensive application is HFM as in our case.

      can you please share your experience for those parameters ? is it possible to improve consolidation speed by focusing more on those parameters or else ?

      Best Regards,
        • 1. Re: Consolidation performance

          Not sure how many users you have working on your system and how much data you are consolidating, but 30 GB sounds like a lot for an HFM App. You can drastically affect speed depending on an App's rules, and in my experience, you can get a much bigger performance improvement by improving the rules than by optimizing those settings.

          1 person found this helpful
          • 2. Re: Consolidation performance



            I am afraid you have data explosion at HFM which is caused by defects in rule file.


            For example, if any rules writes like


            Hs.Exp A#10000 = A#120000.I#[ICP Top].C1#ABC.C2#ABC.C3#ABC.C4#ABC


            If A#120000.I#[ICP Top].C1#ABC.C2#ABC.C3#ABC.C4#ABC = 100, then this 100 will copied to all the available intersection for account 10000.


            However, you may only need to copy this value to one intersection for account 10000.


            This is the most possible reason that you have so many data at HFM.




            James Zhang

            1 person found this helpful