One of the things I enjoy most is diagnosing storage latency. I honestly couldn’t tell you why I enjoy it so much; but, I just do. When it comes to storage in Windows one of the best things you can do is capture an ETL trace for the STORPORT driver.
Databases are platforms that are designed to securely store and retrieve your data. Perhaps that’s why they’re called a data “base”? So if your data is in a base, you’d want to lay it out in some logic way.
Maintaining a database is an important job of the DBA role. One of the many maintenance tasks is ensuring that the disk does not fill up and your files are able to grow over time.
I’ve been testing the new Temporal Tables feature over the past day to see about using it in one of our production databases. It’s a neat feature that honestly adds a boat load of possibility around logging.
In my testing I noticed that user created tables seem to store the rows over quite a bit more pages. User created history tables were nearly double the size of an auto generated one. If you’re currently using the feature or plan to use it in the near future, you’ll want to think about this storage issue before you implement.
I’ve been using Microsoft’s cloud database for some time now. I’ve had a few customers with various performance problems and thought I’d take a moment to highlight the interesting behavior you may run into as well.
I was recently doing some work on my Windows 10 desktop and placed a drive on one of the slower internal disks. It’s old. It’s 7200 RPM. It’s just plain slow. All that’s fine and good as I’m not doing any production workloads from this development machine.
I thought that until one query was taking exceptionally long to complete. I pulled up Resource Monitor to check CPU and storage pressure. Everything looked just fine. In fact, the D: drive wasn’t even listed as having any IO latency.Read More »