In this blog post, we will pick up where we left of in the first one, fixing stuff that did not work. What flummoxed me in the first post was the fact was that the Euclidean distances between articles which were related to each other was greater than the distances between ones that were not related to each other. What I was doing wrong was that, instead considering the whole dataset I limited it only to data from one subset belonging to a particular item and tried to fix its point in space relative to the different set that was not connected to this one in anyway.
I tried to find out if it was easy to find posts similar to one another. If I was successful in doing this then, extending this to find similar posts on various blogs I follow would be the next logical step.
Assuming you know your avg.daily growth rate using a method of your choosing; the numbers for 3 months down the line can be calculated using: (avg.daily growth*90)*(1+x)-(existing free space in DB)-(existing free space on disk) Where, x is a fraction…
Tagged with: Capacity planning
Posted in general
In the final post of the series, we will look at some of the common problems that you might encounter when reading CSV files, and how to circumvent these problems. As usual, first we need to get a list of counters so…
Before I moved to a new topic; I thought, I would give some background on why I wrote “Working with perfmon csv logs in PowerShell – Part 1“. When you are working in a place where the PerfMon logs are collected religiously and always…
In this post we explore how to quickly summarize data from perfmon CSV logs using powershell. The problem can be broken down into three different steps:
1. Identify the columns you want to select
2. Restrict the selection of rows to some convenient size using either date or other selection criteria
3. Measure your selection for min, max and average properties.