Condusiv Technologies Blog

Condusiv Technologies Blog

Blogging @Condusiv

The Condusiv blog shares insight into the issues surrounding system and application performance—and how I/O optimization software is breaking new ground in solving those issues.

Flash Cache Technology: ExpressCache

by Michael 13. January 2011 04:35

ExpressCache is a new, presently shipping, OEM product that we announced at the CES trade show last week.

ExpressCache is a software storage performance solution that requires both an HDD and an SSD (smaller affordables drives are all that is needed). In short, ExpressCache effectively combines the two drives, greatly increasing performance (typically by a factor of 2 or 3) by caching the frequently used data on the SSD!

More Info:

You can read more about ExpressCache here.

Future possibilities:

There are plans to broaden the supported platforms and potentially expand the availability of the product beyond OEMs. The time frame for that is undecided, but it may be later this year. We'll certainly announce any such events on our website and on the blog, so stay tuned.

Tags:

SSD, Solid State, Flash

Diskeeper 2011 outmuscles fragmentation

by Colleen Toumayan 12. January 2011 09:29

Bruce Pechman the "Muscleman of Technology" stopped by out booth at a CES media event. We talked a bit about the upcoming new Diskeeper release and he was kind enough to express his enthusiasm to us in writing so we could publish it:

As a TV Technology Journalist, high-end computer enthusiast, hard core gamer, and a person who is consistently buying the fastest computer money can buy, I need my computers to have super fast performance without any bottle necks every minute of every day.

From expensive SSD’s, to traditional rotational Hard Disk Drives—in every combination you can think of, my computers always run flawlessly and speedily thanks to an awesome software program called Diskeeper!

Right now my primary computer is the “Maingear Shift” with an Intel i980 Extreme Processor overclocked to 4.3 Ghz, it’s almost $6,000 and I can’t tolerate any slowdowns from my dual Western Digital 10,000 RPM VelociRaptors hard drives.

The facts are really quite simple. All types of computers will experience disk fragmentation and it can definitely worsen over time if nothing is done to prevent it. Disk fragmentation is the culprit for many annoying computer symptoms such as slow application performance, long boot times, unexplained system slow downs, crashes, etc. Diskeeper prevents these issues from cropping up.I have been religiously using Diskeeper over the years so my computers can realize the full benefits of system performance and reliability...and I can’t wait to install the new Diskeeper 2011—just set it and “forget about it”!

Whether I’m on deadline, overclocking an Intel Extreme processor, or playing Mafia II with every performance benchmark set to maximum settings, Diskeeper is a must install on every computer I own. 

Bruce Pechman, The Muscleman of Technology (www.MrBicep.com) America’s Best-Built Technology & Fitness Television Personality Muscleman of Technology, LLC Bruce Pechman, The Muscleman of Technology® (Muscleman of Technology, LLC), is a Consumer Technology and Fitness Personality appearing regularly on “Good Morning San Diego” (KUSI-TV) and the KTLA Morning Show in Hollywood. Bruce’s moniker reflects his 20 years of industry technology expertise, and 30 years of fitness training. He has made over 250 live TV appearances on major network television. Visit him at www.mrbicep.com 

Tags:

Defrag | Diskeeper | SSD, Solid State, Flash

Inside SSDs 101

by Michael 31. December 2010 06:08

We have numerous partners and alliances in the  Solid State drive (SSD) space that we interact with regularly. Conversations in related meetings with those allies continues to revolve around the same issue, overcoming performance bottle necks at the storage level. In attacking this problem to get higher performance for things like boot times, application load times, etc., the industry has turned to flash memory, otherwise referred to SSDs (we'll also be announcing a brand new SSD technology in the next few weeks).

The following may be well known to those highly knowlegdable in SSDs, but hopefully helps others less versed in their design.

High end SSDs have proven to yield some very impressive read times, well over double a typical SATA hard disk drive in some cases.

Here are some example published speeds from a few manufacturers/models:

Seagate 7200.12 HDD 500GB, 750GB, and 1TB family

Read/Write speeds (outer tracks, empty drive): 125MB/sec sustained, 160MB/sec peak

Intel X25-M 80GB SSD (MLC)

Sequential Access - Read: up to 250MB/s

Sequential Access - Write: up to 70MB/s

Intel X25-M 120GB SSD (MLC)

Sequential Access - Read: up to 250MB/s

Sequential Access - Write: up to 100MB/s

Intel X25-E 32GB SSD (SLC)

Sequential Access - Read: up to 250 MB/s

Sequential Access - Write: up to 170 MB/s

One of the main reasons for such fast read times is the lack of “seek time” that an SSD has to find and retrieve a piece of data versus a hard drive.  Simply put, a hard drive has to move a magnetic head connected to an arm over a track on a platter and then through various means find the data requested and read or write something.

Now you have to admit, a hard drive does this quite well and very fast, considering the physics involved.

On the other hand an SSD sends an electrical pulse to read the data which is much faster in comparison; give or take on the order of magnitude, double on higher end SSDs. The lack of a moving part cuts the time down considerably.  

Now, writing data to SSDs is a whole other story which leads us down a bit of a rabbit hole so to speak and the main subject of this blog.

SSD DNA

To start with, let’s look at what an SSD is:

Note, I have borrowed some photos and descriptions from the following site: www.popsci.com/category/tags/ssd.

Firstly you simply have a small piece of silicon with a whole lot of transistors that look like the following:

  

Each transistor is 1000 times thinner than a human hair. In essence, each transistor either holds an electrical charge or it doesn’t.  In the case of SSDs, a transistor that is charged equals the value 0 and the ones that are not equal a value of 1. That is the extent of it. 

In the above photo the transistor that is charged has a circle around the “e” which stands for electrons representing the charge.   

Now, to read what’s inside these transistors an electrical pulse is sent to them and by reading the signal from the pulse sent it’s able to tell which are charged and which are not. All in all this is a pretty fast operation. Even writing to a transistor for the first time is pretty fast as it’s empty with no charge to begin with. 

BUT…… what happens when you have to write to an area of the SSD that has already been written to?  Here in lays the problem and the unfortunate “Achilles heel” for SSDs. With a hard drive you can just simply flip the bit in place and call it a day.  In fact this is a faster operation on a hard drive then it is on an SSD, over the long run. Now aside from the extra step it takes to erase an SSD it gets a lot worse. But in order to understand this we need to view how data is laid out on an SSD:

The smallest unit is a single transistor, also known as a cell. Imagine the dot shown is a top view of a single transistor magnified a gazillion times.

This single transister holds a single bit of data, i.e. a 1 or a 0.  The next unit size up is a called a page and holds 4KB of data.   To put this in perspective, there are 8,192 bits in one kilobyte so each page contains 32,768 transistors.   A page is the smallest unit that can be written to on the SSD.  This means that even if the data you are writing is only 1,500 bits in size, it will use up the entire 4KB of space and make it unusable for writing other data.  You only get to write once to it until its been erased for reuse again.

In fact, to update the data within this page, the data would have to be copied first, updated and rewritten to a new page leaving the old page unusable till it’s been erased.  The controller has to earmark it for clean up so it can be used again later.

Now, even though this page has been earmarked for being erased, it may not get erased for some time because it’s actually part of a bigger group of data called a “block”.  A block is a group of pages as illustrated below:

                  

 

The number of pages that make up a block may vary from one SSD model to another and can be very large (in the megabytes).  On one such SSD we have tested on, this happens to be 128k, which would be a group 32 pages.  This data block (32 pages) is actually what’s called an “erase block”.  An SSD only erases one data “block” at a time.  So back to our example of this page with old data in it: theoretically this page could sit around a while before this block of data is erased by the SSD.  It’s plausible this could have in some cases security of data issues but that’s a subject for research and test. 

Now, when data is updated in a page as discussed earlier it has to be moved to another location unbeknownst to the file system and internally mapped by the controller on the device to keep track of it as illustrated below:

There is obviously overhead associated with this. So a page, for argument sake, could have three states of existence:

Ready to be used (it’s erased or been erased thus all transistors have been set to 1) as in the clear little boxes above;

Used -- as in the blue boxes above;

Dirty (containing invalid data that needs to be erased) as in the black boxes above.  

The overhead in handling dirty data is huge, and referred to as garbage collection.  Garbage collection is the process of moving good data out of areas where old data exists in order to erase the old data in order to reclaim pages that can be written to again.

Doing a lot of this garbage collection activity creates a lot of activity called “write amplification”.

This is the disease that SSDs are plagued with which kills their write performance, particularly as the device fills up.  It’s also what shortens the life span of the device.

The following is a definition or description from Wikipedia that I think is pretty good:

“Write amplification (WA) is a phenomenon associated with Flash memory and solid-state drives (SSDs). Because Flash memory must be erased before it can be rewritten, the process to perform these operations results in moving (or rewriting) user data and metadata more than once. This multiplying effect increases the number of writes required over the life of the SSD which shortens the time it can reliably operate. The increased writes also consume bandwidth to the Flash memory which mainly reduces random write performance to the SSD. Many factors will affect the write amplification of an SSD, some can be controlled by the user and some are a direct result of the data written to and usage of the SSD.”

Now, there is a comment that is made in the above description that is interesting, and could lead to read performance in some cases getting degraded, which is the “increased writes” consuming bandwidth to the Flash memory interrupting a read operation.

Now, I don’t say this as a fact but rather postulating whether or not reads are affected. 

The overhead in writes required by the Windows NTFS file system to do just one IO could be considered extreme from the SSD’s point of view.  To create and write one bit of information, such as a notepad doc with a single number in it, requires an update to the MFT file by creating a record of the operation, update to a directory file and any other metadata files such as journal files that are keeping track of operations at the time. The point is, for every one write of user data there are several writes occurring to keep track of it. 

Current file systems were designed based on hard disk drive characteristics, not SSDs.  From the SSD’s point a view, NTFS writes with wild abandon. This puts a lot of overhead on the SSD controllers with a lot of overhead on data mapping and housekeeping at the controller level of the SSD which hard drives don’t have to worry about much.  A hard drive typically only has to re-map data when it has a bad sector on it as an example.

NTFS file system thinks for example that a file should be split in two (split IOs) because it thinks it doesn’t have a contiguous free space for the file that is being updated.  Yet the SSD may have, during its garbage collection process created a space or remapped data clusters but the operating system doesn’t know this and vice versa. 

The current TRIM functionality is supposed to help SSDs within Windows 7 but it's far from being a panacea to the write amplification issue. 

Different types of SSD

SLC stands for “single level cell”

MLC stands for “multi level cell”

TLC stands for “tender loving care". okay not really, just checking to make sure you are paying attention. It really stands for" tri level cell”; really. 

SLC is faster than MLC and TLC. 

The design of an MLC and an SLC are pretty much the same.  The difference is an MLC is able to put more than one value in a single transistor (referred to as a cell) by layering the data within the cell.  Typically two or more bits are able to be placed in a single cell with MLC versus one bit in SLC. 

So MLC is able to contain twice as much data versus SLC.  That’s the plus side.  The down side is MLC in reading and writing data to a single cell has to be very precise and has a lot of overhead logic wise, so MLC ends up slower than SLC due to the preciseness that is required to determine if two values exist and what those values are in a single cell. Also the life cycle of MLC becomes 10x shorter than SLC.  The following is a great white paper that describes the differences very well and how voltages are used to read the values, etc. http://www.supertalent.com/datasheets/SLC_vs_MLC%20whitepaper.pdf

The difference between TLC and MLC is NOT more transistors.  L stands for "level", referring to voltage level to a transistor, not multiple levels or numbers of transistors.  Again the above link along with a few other sites lays this out fairly well. 

The difference between one flash memory or SSD is not so much the chip itself but the supporting controller and its design.  This includes creating multiple data channels, on-board Raid and other fancy and expensive solutions.  The other differentiator is software within the controller managing all of the mapping of data, moving data around, etc.  Sounds like the age-old fragmentation problem to me again; just at a slightly different level.

Diskeeper Corporation Celebrates Its 30th Year With Industry Firsts By Offering Savings of Up to 30 Percent

by Colleen Toumayan 17. June 2010 10:30

Solution Helps Organizations Increase Performance, Lengthen Equipment Lifecycles, Enhance Efficiency and Reduce Energy Usage 

Diskeeper Corporation, innovators in performance and reliability technologies, today announced that it is commemorating  its 30th year of pioneering breakthrough technologies with more than a dozen industry firsts by offering discounts of up to 30 percent on all Diskeeper Corporation volume licenses.  

Company Highlights: 
  • With more than 38 million licenses sold Diskeeper Corporation supports customers worldwide including more than 90 percent of Fortune 500 enterprises, and nearly 70 percent of the Forbes Global 1000, as well as thousands of enterprises, government agencies, independent software vendors (ISVs) and original equipment manufacturers (OEMs).
 
  • CIOs, IT Managers and System Administrators around the world rely on Diskeeper ® performance software to increase computing speeds, reduce system freezes and crashes, improve backup times, lower resource usage, protect data and shorten boot times.
 
  • Every day, Diskeeper Corporation solutions prevent more than 12.5 billion fragments from thrashing hard drives, providing unparalleled performance and reliability for laptops, desktops and servers.
 
  • Diskeeper Corporation’s industry first Undelete® real time data protection solution saves organizations tremendous amounts of time and money by guarding against intentional and unintentional data loss, protecting all deleted files and allowing instant file recovery with just a few mouse clicks.
 Industry Firsts: 
  • 1986: Diskeeper performance software released as the first online automatic defragmenter, which quickly became the best-selling third-party product for OpenVMS operating systems.
  • 1995: Diskeeper is the first defragmenter solution certified for Microsoft Windows, starting a long-running verification process that maintains code reliability.
  • 1995: Diskeeper Corporation partners with Microsoft and co-created APIs that were released with NT 4.0 in 1996.
  • 1998 Network Undelete 1.0 unveiled as the first complete real-time file protection technology for Windows servers.
  • 2003: Diskeeper 8.0 is introduced as a breakthrough approach to help optimize terabyte-sized drives with Terabyte Volume Engine™ technology.
  • 2005: I-FAAST® intelligent file access acceleration sequencing technology introduced that accelerates access to most used files.
  • 2006: InvisiTasking® technology revolutionizes background processing with zero overhead.
  • 2008: Diskeeper releases HyperFast® solid state drive optimizer for PCs.
  • 2009: Diskeeper Corporation introduces V-locity™ virtual platform disk optimizer.
  • 2009: Diskeeper Corporation releases IntelliWriteÔ technology, the first fragmentation prevention technology.
  • 2010: The release of HyperBoot™ boot-time optimization software, which accelerates full computer start up and boots a PC directly into Windows.
  

Tags:

Defrag | Diskeeper | Press Release | SSD, Solid State, Flash

Windows IT Pro Webinar - Should I defrag my SAN?

by Michael 4. May 2010 04:59

Later this month (Thursday, May 27, 2010 at 12:00pm Eastern / 9:00am Pacific), IT Analyst David Chernicoff and I will co-host a webinar covering the benefits and caveats of defragmenting SAN attached servers.

Here is the abstract: 

As storage technologies have become more advanced there is a tendency for storage administrators to believe that the hardware is handling all of their data maintenance needs, keeping their files optimized in the best possible way for top performance and availability. The reality is that hardware solutions alone aren't the most efficient way to keep your critical data stored in an optimal fashion. With this webinar we will give you the information you need to understand how your data is actually being handled and what you can do to improve the performance and optimization of your Windows Server storage.

You can register here: www.windowsitpro.com/go/seminars/diskeeper/windows_san 

We also have two webinars planned for June on the topic of virtualization. One jointly with Redmond Mag, and the other with Microsoft. I'll post registration links on this blog when they are available. Lastly, we'll have a reprise of the SSD webinar we did with Microsoft sometime in July - this time we'll host and Microsoft will be our guest.

Tags:

SAN | SSD, Solid State, Flash | virtualization

Month List

Calendar

<<  September 2017  >>
MoTuWeThFrSaSu
28293031123
45678910
11121314151617
18192021222324
2526272829301
2345678

View posts in large calendar