optimizationinessbase-140305163244-phpapp01

6
OPTIMIZATION IN ESSBASE: Application Performance Optimization can be done by the followin techni!"e# 1. Designing of The Outline using Hour Glass Model 2. Defragmentation 3. Restructuring 4. Compression Techniues !. Cache "ettings #. $ntelligent Calculation %. &ncommitted 'ccess (. Data )oad Optimi*ation $e#inin of The O"tline "#in %o"r &la## Model: Outline should +e designed in such a ,a- that dimensions are placed in the follo,ing order largest dens smallest dense/ smallest sparse to largest sparse follo,ed +- 'ttri+ute Dimensio &sing hourglass model impro0es 1 of calculation erformance of the cu+e. $eframentation: ragmentation is caused due to the follo,ing 1. reuent Data )oad 2. reuent Retrie0al 3. reuent Calculation 5e can chec6 ,hether the cu+e is fragmented or not +- seeing its '0erage Clustering Ratio in the properties. The Optimum clustering 0alue is 1/ $f the a0 clustering ratio is less than 1/ then the cu+e is fragmented ,hich degrades the performance of the cu+e. There are ' way# of doin $eframentation: 1. 78port Data of the application in to te8t 9le/ then clear data and reload t data using te8t 9le ,ithout using Rules 9le. 2. &sing M':) Command; Ma8l<'lter Data+ase 'ppname.D= name orce Restructure

description

optimization in essbase

Transcript of optimizationinessbase-140305163244-phpapp01

OPTIMIZATION IN ESSBASE:

Application Performance Optimization can be done by the following techniques1. Designing of The Outline using Hour Glass Model2. Defragmentation3. Restructuring4. Compression Techniques5. Cache Settings6. Intelligent Calculation7. Uncommitted Access8. Data Load OptimizationDesigning of The Outline using Hour Glass Model: Outline should be designed in such a way that dimensions are placed in the following order - largest dense to smallest dense, smallest sparse to largest sparse followed by Attribute Dimensions. Using hourglass model improves 10% of calculation Performance of the cube.Defragmentation: Fragmentation is caused due to the following1. Frequent Data Load2. Frequent Retrieval3. Frequent CalculationWe can check whether the cube is fragmented or not by seeing its Average Clustering Ratio in the properties. The Optimum clustering value is 1, If the average clustering ratio is less than 1, then the cube is fragmented which degrades the performance of the cube.There are 3 ways of doing Defragmentation:1. Export Data of the application in to text file, then clear data and reload the data using text file without using Rules file.2. Using MAXL Command:Maxl>Alter Database Appname.DB name Force Restructure3. Add and Delete One Dummy Member in the Dense Dimension .Restructuring: There are 3 types of Restructure.1. Outline Restructure2. Sparse Restructure3. Dense Restructure/Full RestructureOutline Restructure: When we rename any member or add Alias to any member then outline Restructure would Happen..OTL file is converted to .OTN which in turn converts in to .OTL again..OTN file is a temp file deleted by default after restructureDense Restructure(Full Restructure): If a member of a dense dimension is moved, deleted, or added, Essbase restructures the blocks in the data files and creates new data files. When Essbase restructures the data blocks, it regenerates the index automatically so that index entries point to the new data blocks. Empty blocks are not removed. Essbase marks all restructured blocks as dirty, so after a dense restructure you must recalculate the database. Dense Restructuring, the most time-consuming of the restructures, can take a long time to complete for large databases.

Sparse Restructure: If a member of a sparse dimension is moved, deleted, or added, Essbase restructures the index and creates new index files. Restructuring the index is relatively fast; the time required depends on the index size.Compression Techniques: There are 4 types of Compressions. They are1. Bitmap Compression2. RLE Run length Encoding3. ZLIB4. No Compression.Caches: There are 5 types of caches.1. Index cache2. Data Cache3. Data File Cache4. Calculator Cache5. Dynamic Calculator CacheIndex Cache: Index Cache is a buffer in a memory that holds Index Files (.IND). Index cache should be set equal to the size of the index file.Note: Restart the database in order to make the new cache settings come in to effect.Data Cache: Data cache is a buffer in a memory that holds Uncompressed Data Blocks.Data cache should be 12.5% of the PAG file memory, by default it is set to 3MB.Data File Cache: Data file cache is a buffer in memory that holds compressed data blocks.Size of the Data file cache should be size of the PAG File memory. It is set to 32MB by default. Max. Size for data file cache is is 2GBWe can use only either Data cache/ Data file cache most of the developers prefer Data cache in Real time.Calculator Cache: It is basically used to improve the performance of calculation.WE set the calculator cache in calculation scripts.Set cache High|Low|Off; ----- command used in calc scripts to set the cache.We set cache value for calculator cache in Essbase.cfg file.We need to restart the server to make the changes in calculator cache after setting it in config file.Dynamic Calculator Cache: The dynamic calculator cache is a buffer in memory that Essbase uses to store all of the blocks needed for a calculation of a Dynamic Calc member in a dense dimension (for example, for a query).Intelligent Calculation: Whenever the Block is created for the 1st time Essbase would treat it as Dirty Block. When we run Calc all/Calc dim Essbase would calculate and mark all blocks as Clean blocks. Subsequently, when we change value in any block the block is marked as Dirty block. when we run calc scripts again only dirty blocks are calculated it is known as Intelligent Calculation.By default Intelligent calculation is ON. To turn off the Intelligent Calculation use command SET Update Calc OFF; in scripts .Uncommitted Access: Under uncommitted access, Essbase locks blocks for write access until Essbase finishes updating the block. Under committed access, Essbase holds locks until a transaction completes. With uncommitted access, blocks are released more frequently than with committed access. The Essbase performance is better if we set uncommitted access. Besides, parallel calculation only works with uncommitted access.

Data Load Optimization: Data load optimization can be achieved by the following.1. Always load the data from the Server than file system.2. The data should be at last after the combinations.3. Should use #MI instead of 0s. If we use 0 uses 8 bytes of memory for each cell.4. Restrict max Decimal Points to 3 -- 1.2345. Data should be loaded in the form of Inverted Hourglass Model.(Largest sparse to Smallest Sparse followed by smallest Dense to Largest Dense data)6. Always Pre-Aggregate data before loading data in to Database.DL Threads write (4/8): Used for Parallel Data loads. Loads 4 records at a time for 32-Bit system and 8 records for 64-Bit system.By default Essbase Loads data Record by Record which would consume more time resulting in consuming huge time for data loads.Optimization Techniques in Essbase The best technique to make large data loads faster is to have the optimal order of dimensions in source file, and to sort this optimally, order the fields in your source file (or SQLstatement) by having hourglass dimension order, you data file should have dimensions listed from the bottom dimension upwards. Your dense dimensions should always be first, and ifyou have multiple data columns these should be dense dimension members. This will cause blocks to be created and filled with data in sequence, making the data load faster and thecube less fragmented.As a part of Optimization we need to re-order the dimensions as follows Large members Dense dimension Small members Dense dimension Small members Sparse dimension Large members Sparse dimension Attribute dimensions.

Calculation order of the dimensions. Dimension tagged accounts if it is dense. Dense dimensions in outline or CALC DIM statement order. Dimensions tagged as Accounts if it is sparse. Sparse dimensions in outline order or CALC DIM statement order. Two-pass calculations on members in the Accounts tagged dimension.

Here are some more optimization techniques used in Essbase For data loading:

Grouping Sparse Member Combinations Positioning Data in the Same Order As the Outline Loading from the Essbase OLAP Server Making the Data Source As Small As Possible Making Source Fields As Small As Possible Managing Parallel Data Load Processing

For Calculation: Using Parallel Calculation Using Formulas Managing Caches to Improve Performance Using Two-Pass Calculation Aggregating #MISSING Values Removing #MISSSING BlocksFew Optimization Techniques in Essbase

With the essential features available in Essbase you can load the huge data to the Essbase cubes, Run the reports and you can perform the complex calculations also,As you keep on adding the different features to your application the performance will get reduce. As i said Essbase came up with different features along with the different performance tuning techniques which makes the application best optimized.The optimization can be done at many places such as

Outline Optimization:

1) Arrange the dimension in"Hour Glass Model" The Outline should starts with dense dimension with highest stored members and it keep going till the dense dimension with least stored members and then starts with sparse dimension with least stored members and it keep going till the sparse dimension with highest stored members.2) Use the member storage properties efficiently. If the dimension is to just host the different types of data such as scenarios, here there is no point in rolling up the lower values to higher level, in this situation you can tag the dimension as "Label Only" and assign the no consolidation operator to the members under it. Some calculations really not required to stored the results in database at this point of time tag the concern members with "Dynamic Calc" property.

Data Load Optimization:1) In data file, the fields should starts with sparse dimension members and then dense dimension members and then the data field.2) If the same field is repeating in all the records in the data file, then try to ignore that field from fetching itself and keep that member in the "Header Definition", why means to save the buffer memory and it will increase data load process.

Report Script Optimization:1) In the report script first specify the sparse dimensions and then dense dimensions, why means :Sparse dimension creates the data blocks within which the data cells are available, so specifying the dense first does not make sense. So to speed up the process specify the data blocks first(Sparse dimension) and then data cells (Dense dimensions).2) The dimensions which are not required to display in the report put them in thepage.3) Use the special commands to increase the report performanceSUPMISSINGROWS :To Suppress the data missing rows.SUPHEADING :To Suppress the headings.SUPBRACKETS:To Suppress the brackets around the negative values.SUPEMPTYROWS:To Suppress the empty rows.

Calculation Script Optimization:1) Use the set commands to increase the calculation performance. SET MSG SUMMARY: Set the message level to summary. SET AGGMISSG ON :To avoid the aggregation of #missing values. SET CACHE HIGH :To increase the bugger size. SET NOTICE LOW :To set the notices to low.2) Perform the calculations on only required part of database usingFIXcommand.