Here's another idea: If you can join the pageview and memory by timestamp values, you can create a table like this
Page A | Page B | Page C | Page D | Page E | .... | Memory_use
The value for each page column can be a bit [0,1], indicating that the page was requested or not, or the number of pages depending on your data. In the column Memory_use you can have the appropriate proportions of the memory load or count in MB. Thus, Memory_use can be considered as a dependent variable, and pages as explanatory. Thus, you can choose a suitable (depending on the form of the dependent variable) generalized linear model for this data set. The results of this analysis will give you an idea of ββthe following
-What pages significantly affect memory usage
- the degree to which each page contributes to the load (by its coefficient in the model)
- The possibility that other factors, not measured, play a significant role in the memory load (overdispersion), in the worst case, that all predictor variables may turn out to be insignificant.
source share