Posted in Uncategorized on February 23, 2021 by RZGiampaoli
Hey guys, how are you? Continuing the Essbase Statistics DW: How to automatically administrate Essbase using ODI series (Part 9 Here), today we’ll talk about the analyzes we can do with all the data we have so far.
With all this information in our DW we can do a lot of things. We could automatically set the caches using ODI and Maxl or Esscmd but change settings in Essbase is more complicated than just an update.
We need to restart the applications, we need to take in consideration the design of each cube, we need the feeling and knowledge of the administrators, but can be done.
In fact, we can create thresholds to see if the cache is 20% off what it supposed to be and then apply the changes at night. We can also start to record the sessions (be aware that the tables get very big very quick), and analyze the retrieve/calculation times to see if the performance is getting better or not and them send a alert in case it was not after a automatic chance or something like that.
But what we can really do is provide precise information about each cube. We send a daily email for each administrator saying the actual stats for each cube, the actual setting of each cache, the recommended size, and we also send a table with more information about that application.
Attached we can add an excel with the raw data for further analysis.
This helps a lot the administrators to have a better understanding of our cubes and to decide and plan better maintenance strategies.
Of course, we can have a dashboard (and this would be the best approach) with all the statistics there showing the growing of the cube, projections, alerts of when the cubes will reach the maximum cache size, well, anything that could be needed.
The important thing here is to realize how useful this DW would be for keep our Essbase cubes healthy. And keep in mind that is a lot of other stats that can be gathered from Essbase that can do a lot for us.
For example, we can get the dimension settings and number of members and use ODI (or any language like java) to manipulate the outline and test the best Sparse/Dense settings for our cubes.
I hope you guys enjoy our series, be safe and see you soon.
Now its time to glue it all together with ODI. ODI is great here because it can work with different technologies without any effort. In our case, we will run the Java codes that we displayed in this series using Java BeanShell Technology.
Although ODI is great to execute any kind of technology code out there, it does not have a good way for you to write and debug your code while you are developing it. So, I always prefer to create the code in an external Java IDE (like Eclipse), test it and copy the “Main” portion of it into an ODI Procedure.
Another advantage to use ODI is that we can get the connection information from topology itself and reuse everywhere within the package using the command on source/target technique, where you define the connection in the command on source tab and get the information in the command on target tab.
To glue all together in ODI is very simple. We may have all the java codes in one procedure with as many steps as we want (depending in what kind of statistics we want to get, in our case, three). Then we may have another procedure that will hold our pivot queries that we use to transform and load the data into our DW tables. Finally, we may even create our own metrics based on the knowledge that we have from Essbase. Down below is one example of metrics that we may retrieve from the stats that we just gathered.
When we put it all in an ODI package, it will look like this. In this example, we also added a send email component just to inform the users about the job completion.
That’s it folks! Next post will be the last one of this series. Stay tuned!
Before we glue it all together in ODI, let us organize the data that we just got from Essbase. In the last post we saw that we gathered all the information we needed inside a generic stage table. Although a generic table is great from an extract perspective, we may decide to split, organize, and load this information in some kind of “historical DW tables” which will help us to analyze the data better over time.
For this series of posts, we have 3 different kind of information: DB Statistics, File Statistics and Outline Statistics. Their structure are very different, so we will generate one historical table for each kind of statistic.
Since we created a generic table to hold all extracted information in rows, now we need to PIVOT the data into columns and load it into the historical tables. This can be easily achieved with a SQL like the one below:
First, we define the columns to be pivoted and use a consolidation function on the data column, like (SUM, AVG, MIN, MAX, COUNT…). Then we specify the data to be pivoted and this data must be a constant in the “IN” clause. Finally, this data is loaded to the DW table using a Control Append approach, since we want to keep adding the current statistics to a historical table for an analysis over time.
If we follow this approach for all three metrics that we retrieved from Essbase, we will end up with three DW tables, like the ones below:
Hi all, let us demonstrate some examples of Java codes that you may use to retrieve statistics out of Essbase. Since we want to store those stats in an Oracle table, let us begin with an example of how to connect to an Oracle DB, which is simple, as we just need the DB URL, username, and password. Below is how we may create the connection. We also have a prepare statement that we will use to do inserts in our Oracle table.
Since we have a generic table, we can have just one prepare statement and use it to insert all kind of stats there.
Connecting to Essbase is also easy, the only additional information is that you also need to pass the Provider Service, which may be “Embedded” or by Provider Service. Here we are basically doing a sign in to the provider service and then select the OLAP server that we want to use.
Now let us see the first code to get some cube stats. First, we get all the cubes from all the apps and for each cube we get its properties (which is an array in a Key and a Value format). This information is added to the prepare Statement and executed (inserted) into the DB.
The result will be something like this:
Let us jump to a second example. In Java we can issue any Maxl command using the “IEssMaxlSession” class. The result set contains columns and rows, similar with what we get in EAS. Then we need to loop through the rows and get the columns that we need. This information is also added to the prepare Statement and executed (inserted) into the DB.
The result will be something like below.
The last example is how we can get stats from the Essbase outline. We can get member information using IEssMemberSelection with a custom query or find the member directly in the outline using find member. The results contain a set of members that we may loop and analyze its properties. In this case we decided to categorize the results by its storage type.
Again, the result will be something like the one below.
That is it for today. Next post I will show you how you can glue it all together with ODI. See you soon!
Hi all! Let us talk now about how we can automate this stat gathering using Java Essbase API. Java is the key technology here since it can easily connect and manipulate Essbase through its API. It also can connect to Oracle Database to store our results, run OS commands and more all in one single code. Java is also great since it may be easily deployed to ODI using Procedures and scheduled using ODI Operator. All in all, combining ODI and Java code creates a powerful and seamlessly integration going beyond the database boundaries.
Let us begin with some Java Essbase API basics. The main goal is to develop one single code that will connect in Essbase, retrieve the statistics information and load that in the Oracle database.
Essbase API is very similar with what we see in EAS, in the sense that the structure of the classes follows the same architecture as in a Essbase server (Server->App->Cube->Otl), which makes it easy to find out what you are looking for when looking at the API documentation.
Since we will store this information in an Oracle table, we will also need to know a little bit about Oracle Java API, but luckily this one is straight forward.
With those two sets of APIs, we are good retrieve all the information that we need from Essbase. Each stat has its own number of columns and metrics, so if we create one table for each kind of structure it will be very tricky to maintain and harder to create any kind of generic code. The best way to extract information is to have just one table where we have the properties in the rows instead of columns, this way we have just one structure for all kind of information, no matter the number of columns that returns and we may create generic code around it. Our final table would look like below.
That is it for this post. Next one I will share some examples on how to connect to Oracle DB, Essbase and how to retrieve some stats out of it. Stay tuned!
Posted in Uncategorized on February 4, 2021 by RZGiampaoli
Hey guys, how are you? Continuing the Essbase Statistics DW: How to automatically administrate Essbase using ODI series (Part 3 Here), today we’ll talk about Dynamic Calculator Cache.
The Dynamic calculator cache basically it is a buffer to store all uncompressed blocks on memory to calculate the dynamic members in the dense dimensions. Expanded Blocks including: Store, Never Share, Dynamic Calc members and dynamic time series members.
This is important because show how everything is related in Essbase. If you set the CALCLOCKBLOCK to 1000 you need to be sure that the data cache can hold 2000 uncompressed blocks. If not doesn’t matter the setting you put, Essbase will put half what fits in the data cache
We also need the size of the expanded blocks is all the store members plus the dynamic calc members and dynamic time series members together plus 8.
We also need the largest expanded block across all database on the machine and the maximum number of expected concurrent users. This can be analyzed by gathering the session information into our DW as well, but for this post it’ll be a constant number.
This information can be acquired by:
EAS
MAXL
ESSCMD
JAVA
Data Cache Right click in a cube edit properties->Storage
query database sample.basic list all file information;
listfiles “” “sample” “basic”;
maxl.execute(“query database ” + s_application + “.” + s_cube + ” list all file information”);
Amount of members Right click in a cube edit properties->Dimensions
A buffer in memory that Essbase uses to store all the blocks needed for calculate a Dynamic Calc member in a dense dimension
To find the optimal size of this cache we need:
CALCLOCKBLOCK size: it is half the number of expanded blocks that fit into the data cache
Expanded Blocks including: Store, Never Share, Dynamic Calc members and dynamic time series members
The largest expanded block across all databases on the machine.
The maximum number of expected concurrent users
Can be analysed by gathering the session info into a table and analyse the patterns but for this presentation is a constant number based in experience
To calculate the Maximum Calculator Cache, we need to multiply:
CALCLOCKBLOCK: (Data Cache in bytes (already calculated) / Size of the expanded Block in bytes) / 2
The largest expanded Block in bytes on the server
The maximum number of expected concurrent Users (Constant)
And that’s it, we talked about the 5 most important caches in Essbase, how do we calculate them and how we get them.
So far, we talked about these caches and how they are calculated:
Cache Type
Description
Index
A buffer in memory that holds index pages (.ind files)
Data file
A buffer in memory that holds compressed data files (.pag files). Essbase allocates memory to the data file cache during data load, calculation, and retrieval operations, as needed. (only when direct I/O is in effect)
Data
A buffer in memory that holds uncompressed data blocks. Essbase allocates memory to the data cache during data load, calculation, and retrieval operations, as needed
Calculator
A buffer in memory that Essbase uses to create and track data blocks during calculation operations
Dynamic calculator
A buffer in memory that Essbase uses to store all of the blocks needed for a calculation of a Dynamic Calc member in a dense dimension
And here we have a resume of all calculations needed to have the exactly amount of cache per application we have:
Cache Type
Description
Index
number of existing Blocks * 112 bytes = the size of database index
Data file
Combined size of all essn.pag files, if possible; otherwise, as large as possible
Data
0.125 * the value of Data File Cache size
Calculator
Bitmap size in bytes * Number of bitmaps: Bitmap size in bytes: Max((member combinations on the bitmap dimensions / 8), 4) Number of bitmaps: Maximum number of dependent parents in the anchoring dimension + 2 constant bitmaps
Dynamic Calculator
C * S * U: C is the value of the appropriate CALCLOCKBLOCK setting. (Data Cache in bytes / Size of the expanded Block S is the size of the largest expanded block across all databases on the machine. Multiply the number of members in each dense dimension together and multiply by 8 bytes U is the maximum number of expected concurrent users
From now on we’ll going to see how we can create a process to extract and handle this information and how can we create a DW and use it to keep our Essbase apps with the right cache all the time.
Posted in Uncategorized on February 3, 2021 by RZGiampaoli
Hey guys, how are you? Continuing the Essbase Statistics DW: How to automatically administrate Essbase using ODI series (Part 2 Here), today we’ll talk about Calculator Cache.
The concept behind the Calculator cache is a little tricky but I think is very nice to understand since it gives us a very good tip about Essbase design. This cache is used to track what Essbase already calculated and what he didn’t yet.
Essbase split the sparse dimensions in 2 groups:
Bitmap dimensions: they are all the sparse dimension that fits inside the bitmap
You can think about this bitmap as a matrix that will hold all members from one or more sparse dimension, depending of the available memory. Each member occupies 1 bit of memory. (it’s like a check that says if the member was calculated or not)
Anchoring dimensions: For the rest of the sparse dimensions that didn’t fit in the bitmap. The anchoring dimension will define the number of bitmaps that Essbase can create at same time, if there’s enough memory.
The calculator cache controls the size of the bitmap, what means, the number of sparse dimensions that can fit in the bitmap and the amount of anchoring dimensions depends of the size calculator cache
But how it works? Essbase will start from the first sparse dimension in the outline and will try to fit as much dimensions possible inside the bitmap until he cannot put the next sparse dimensions completely inside the bitmap.
All the remain sparse dimension becomes anchoring dimensions.
The number of bitmaps that can be created at same time is determined by the maximum number of dependent parents in the anchoring dimensions.
Dependent parent is the number of parents that is above a leaf member. We need to find the member that has the biggest number of levels above him, meaning, the max number of dependent parents is the same as the number of Levels in a dimension
If the leaf member with most parents above him has 3 parents the level of that dimension will also be 3, then for dependent parents we just need to ask for the level of the dimensions.
The size of the bitmap and the number of anchoring dimensions depends of the size of the calculator cache.
What that means? Let’s consider this outline for example:
Sparse Dims
# of Members
Dependent Parents
S1
10
NA
S2
10
NA
S3
20
NA
S4
30
NA
S5
200
3
In this outline we have 5 sparse dimensions, 4 flat ones (S1 to S4) and one with a 3-level hierarchy (S5).
We have 3 options of cache settings that we can choose from:
Single anchoring multiple bitmaps.
Single anchoring single bitmap.
And multiple anchoring single bitmap.
To use the first option, we need to get the first 4 sparse dimensions from the outline and multiply all of them, split the result by 8 to transform it in bytes and this will define the size of the bitmap.
After that we need to define how many bitmaps we will have at same time in memory. For that we need to get the level of the anchoring dimensions and sum 2 on it. (2 is a constant number). This will define the number of bitmaps.
Finally, we just need to multiply the size of the bitmap by the number of bitmaps and this will give us the size of the cache we must set to use option 1.
The other options are a smaller version of the first one. Option 2 we don’t have multiple bitmaps, we just set the anchoring to 1, and option 3 we fit in the bitmap less dimensions.
We can see the calculations for each option below:
Option
Calculator Cache
Single anchoring, multiple bitmaps
(10*10*20*30)/8=7500 3+2=5 7500*5=37500 bytes
Single anchoring, single bitmap
(10*10*20*30)/8=7500 1 7500*1=7500 bytes
Multiple anchoring, single bitmap
(10*10*20)/8=250 1 250*1=250 bytes
The interesting thing here is to see why is important to have our outline designed as hourglass for this cache. If we put the most complex sparse dimension in other position than the last one, we will never be able to have concurrent bitmaps and the size of the bitmap will be huge.
Then after this trick explanation the only information we need to calculate the calculator cache are:
The amount of member in each sparse dimension.
The amount of levels in the biggest sparse dimension.
The biggest sparse dimensions should be placed in the last place on the outline.
All this information can be acquired by:
EAS
MAXL
ESSCMD
JAVA
Amount of members Right click in a cube edit properties->Dimensions
Calculator cache is a buffer in memory that Essbase uses to create and track data blocks during calculation operations
Essbase split the sparse dimensions in 2 groups:
Bitmap dimensions: Any sparse dimensions that can be fit inside the bitmap. Each member combination placed in the bitmap occupies 1 bit of memory, and the entire dimension must fit inside the bitmap.
Anchoring dimensions: The remaining sparse dimensions that do not fit into the bitmap.
Because the calculator cache controls the size of the bitmap, the number of sparse dimensions that can fit in the bitmap and anchoring dimensions depends on the size of the calculator cache
Essbase starts with the first sparse dimension in the outline and fits as many as possible into the bitmap.
Essbase stops the process when it cannot fit another complete sparse dimension into the bitmap.
All the remaining sparse dimensions becomes anchoring dimensions
The number of bitmaps used is determined by the maximum number of dependent parents for any members in the anchoring dimension
The max number of dependent parents is the same as the number of Levels in a dimension
Depending of the cache setting we can have 3 options
Single anchoring multiple bitmaps.
Single anchoring single bitmap.
And multiple anchoring single bitmap.
And to calculate everything we just need 2 things:
The amount of member in each sparse dimension.
The amount of levels in the biggest sparse dimension.
The biggest sparse dimensions should be placed in the last place on the outline.
I hope you guys enjoy it, be safe and see you next time.
Posted in Uncategorized on February 2, 2021 by RZGiampaoli
Hey guys, how are you? Continuing the Essbase Statistics DW: How to automatically administrate Essbase using ODI series (Part 1 Here), today we’ll talk about Data File Cache and Data Cache.
The data file cache is deprecated, and I don’t know if much people use it, but it’s used to define the maximum size of the data cache. The data file cache is used to hold compressed data files (essn.pag) in memory and is used only when direct I/O option is in effect. To find it maximum size we need to sum all essn.pag files, if possible; otherwise, as large as possible.
We can find this information in EAS edit properties, storage or Maxl query database file information or in Esscmd list files or java using Maxl as we can see here:
EAS
MAXL
ESSCMD
JAVA
Right click in a cube edit properties->Storage
query database sample.basic list all file information;
listfiles “” “sample” “basic”;
maxl.execute(“query database ” + s_application + “.” + s_cube + ” list all file information”);
Now let’s talk about the data cache. The name looks almost the same, but his usage is very different. The data cache is used hold uncompressed data blocks in memory. It has a direct relationship with the Data file cache and the recommended size would be 12.5 % of the summing of all essn.pag files (or 0.125 * Data File Cache size. That is why we even talked about the data file cache and calculated its size before.
The same thing that happens with the index cache happens here with the data cache. If we have too much historical data, we may decrease the optimum size to fit the data usage.
We gather this information in the same way as the data cache file:
EAS
MAXL
ESSCMD
JAVA
Right click in a cube edit properties->Storage
query database sample.basic list all file information;
listfiles “” “sample” “basic”;
maxl.execute(“query database ” + s_application + “.” + s_cube + ” list all file information”);
Then to resume what we talked about in this post:
The data file cache is used to hold compressed data files (essn.pag) in memory and is used only when direct I/O option is in effect.
It is calculated by summing all essn.pag files, if possible; otherwise, as large as possible.
The data cache is used hold uncompressed data blocks in memory.
The recommended size would be 12.5 % of the summing of all essn.pag files (or 0.125 * Data File Cache size).
Same case as Index cache, just the working blocks are used (no historical).
I’ll stop here since we have a big one coming and I want to address the Calculator Cache alone in the next post.
Posted in Uncategorized on February 1, 2021 by RZGiampaoli
Hey guys, how are you? Today we’ll going to talk about how to automatically administrate Essbase using ODI.
As you guys know, in order to have good performance in an Essbase cube, we must keep vigilance and follow up its growth and its data movements, so we can distribute caches and adjust the database parameters accordingly. But this is a very difficult task to achieve, since Essbase statistics are not temporal and only tells you the cube statistics is in that specific time frame.
The performance of an Essbase app has a direct relationship with the configuration setting that we set and must be based on the design and usage of each cube. To find out the optimal values for these configurations we need to look at the Essbase statistics.
The problem with this is that the Essbase statistics is a snapshot of the cube at that moment and every time we shut down a cube the run time statistics are cleared, meaning that these statistics are more or less accurate depending of the amount time the cube is online.
That also means that if we look in the statistics after a restart of the cube, we can get the wrong information and we can make a poor setting of caches and other things making the cube slow or even crashing them.
The best thing we can do to make this process more reliable is to create a historical DW with the Essbase statistics information, this way we can create a process to keep improving Essbase setting in a more pragmatically way extract, analyze, change setting and so on.
These are the most important caches that we can set, but we can do all kind of things with the information we’ll gathering from Essbase.
Cache Type
Description
Index
A buffer in memory that holds index pages (.ind files)
Data file
A buffer in memory that holds compressed data files (.pag files). Essbase allocates memory to the data file cache during data load, calculation, and retrieval operations, as needed. (only when direct I/O is in effect)
Data
A buffer in memory that holds uncompressed data blocks. Essbase allocates memory to the data cache during data load, calculation, and retrieval operations, as needed
Calculator
A buffer in memory that Essbase uses to create and track data blocks during calculation operations
Dynamic calculator
A buffer in memory that Essbase uses to store all of the blocks needed for a calculation of a Dynamic Calc member in a dense dimension
Let’s start with the Index cache. Index cache holds in memory the indexes that Essbase uses to find requested data.
Every time we retrieve or calculate any data from Essbase, first Essbase try to find that index in the index cache. If the index is not there, Essbase will find the index in the index files, get the data and store this index in the index cache memory to make it faster for the next time we need the same information. For each index that Essbase store there, it will take 112 bytes of memory.
Them to calculate the maximum size of this cache we just need to multiply the number of existing blocks by 112 bytes and that’s it.
Just to make sure everyone is in the same page: a Block is the combination of all store like members from all dense dimensions. Essbase creates a new Block when a new unique combination of data in the sparse dimensions is inserted/calculated.
This table shows the different ways to get the information we need from Essbase. As you can see, we don’t need to know java/C++ or any language to do extract this information, we can just use maxl, esscmd or even manually using EAS.
EAS
MAXL
ESSCMD
JAVA
Right click in a cube edit properties->Statistics
query database sample.basic get dbstats data_block;
select sample basic; getdbstats;
(IEssCube) cube.getPropertyNames();
One interesting thing about this cache (and almost every cache in Essbase) is that, as I said, it holds in memory the indexes that Essbase used at least once to find the requested data. That means, we will only find in the cache, indexes that we used at least once since the last time the cube starts.
Then, if we have a cube with historical data, the chances of these old data never hit the index are bigger then the new data. We need to take this in consideration when we set this cache.
Historical Data Usage
In this case, 70% of the cube contains historical data that are rarely requested, meaning, we need just 30% of the maximum size for the index cache.
A way to figure out if the cache is well set is looking at the hit ratio in EAS, but we still have the issue that this info reset every time when the cube restarts.
If we are resuming what we just said about Index Cache, we have the follow:
The index Cache is used to hold in memory all used indexes since the application is turned on.
The maximum index cache size is calculated by multiplying the number of existing Blocks * 112 bytes
A Block is the combination of all store like members from all dense dimensions. Essbase creates a new Block when a new unique combination of data in the sparse dimensions is inserted/calculated
The optimal index cache will vary depending of the design and use of the cube.
If the cube has a lot of historical data, the chances of these index never hit the cache are big
For that we can define a historical percentage that we can remove from the maximum index cache to find a more realistic cache.
This historical percentage can be obtained by analyzing the outline and see how many members are in use
That’s it for today. Next post we’ll talk about the Data File Cache. Stay safe and see you soon.