Condusiv Technologies Blog

Condusiv Technologies Blog

Blogging @Condusiv

The Condusiv blog shares insight into the issues surrounding system and application performance—and how I/O optimization software is breaking new ground in solving those issues.

The Tsunami of Data is Swamping IT

by Jim D’Arezzo, CEO 29. September 2017 09:53

There’s a storm brewing and chances are, it’s going to hit your data center or cloud site.

It’s the tsunami of data that is washing over the IT world and it is just going to get worse. That’s because there is an insatiable demand for data: big data, ERP, CRM, BI, EHR, and of course all of that social media and video that is flooding global organizations like a bursting dam.

I’ll admit it, I’m a dinosaur. I got started in the IT industry in the late 1970’s at IBM. So I can rightly say that I’ve seen a lot in the past 40 years. When I started, it was the heyday of the mainframe era. Data processing was still practically a priesthood back then. It wasn’t for everyone. Bill Gates would not proclaim “Information at your fingertips” for at least another dozen years.

Clearly, we’ve come a long way in four decades. The three driving factors – exponentially increased compute power, nearly unlimited storage, and the internet have delivered computing nirvana. Or so it would seem. But with abundance comes challenges. Frankly, we are now awash in data. The tsunami of available information is swamping IT environments worldwide. And even with huge advances in the three driving factors (compute, storage and the internet) IT is experiencing performance bottlenecks on an increasing basis. Recently, we conducted a survey of 1400 IT professionals and fully 27% said they are experiencing performance problems that are causing user complaints and slowdowns.

As usual, our industry is innovating to keep up. Cloud based computing that takes advantage of huge data centers as well as innovation in storage, compute power and connectivity has kept up with much of the demand. However, most of these improvements focus on hardware – all flash storage arrays; 64-core servers; hyper-convergence. But there’s one other performance innovation that is often overlooked: Software.

Software innovation that takes advantage of existing and new hardware capabilities can significantly increase performance without the cost of additional hardware. Even with an all flash storage back-end, software like ours can significantly increase performance and extend the life of the hardware. The cost-benefit can be tremendous. Imagine getting a 50% boost in performance (or even 2-3X) without buying a single additional piece of hardware. CEOs and CFOs love that kind of benefit.  I know, because that is what we hear from our customers every day.

So, before you get swamped by the tsunami of data that’s lapping at your data center, consider a software solution to the problem. It can help you stay afloat without having to buy a new yacht!

 

Tags: , , , , , , , ,

Application Performance | Big Data | Cloud

Microsoft SQL Team Puts V-locity to the Test

by Brian Morin 15. September 2017 09:12

In a testament to Condusiv's longstanding 20+ year relationship with Microsoft® as a Gold Partner and provider of technologies to Microsoft over the years, Condusiv® became the first software vendor awarded the stringent certification of MS-SQL Server I/O Reliability joining a very short list containing the likes of Dell® / EMC®, IBM® and HPE®.

Microsoft developed the SQL Server I/O Reliability Program to ensure the reliability, integrity, and availability of vendor products with SQL Server. The program includes a set of requirements that, when complied with and approved by a Microsoft committee of engineers, ensure the product is fully reliable and highly available for SQL Server systems. The certification applies to SQL Server running on Windows Server 2008R2 and later (the most current 2016 release included).

V-locity® Certified for SQL I/O Reliability and Demonstrates Significant SQL Performance Gains

The program itself does not require performance characteristics of products, but it does require I/O testing to exhibit the reliability and integrity of the product. To that end, the full report links to a summary of before/after performance results from a HammerDB test (the preferred load test to measure MS-SQL performance) on Azure to demonstrate the gains of using V-locity I/O reduction software for SQL Server 2016 on Azure’s Windows Server 2016 Data Center Edition. While transactions per minute increased 28.5% and new orders per minute increased by 28.7%, gains were considered modest by Condusiv’s standards since only a limited amount memory was available to be leveraged by V-locity’s patented DRAM caching engine. The typical V-locity customer sees 50% or better performance improvement to SQL applications. The Azure test system configured by Microsoft did not boost available memory to showcase the full power of what V-locity can do with as little of 2-4GB of memory.

To read the full report CLICK HERE

 

Undelete Can Do That Too?

by Gary Quan 30. August 2017 04:46

You may have already heard countless customers tout the file recovery features of Condusiv’s Undelete® and how IT Pros use it as a recycle bin on file servers for real-time protection, so they don’t have to dig through backups to recover deleted or overwritten files. Although this is Undelete’s primary function, Undelete provides more than just this. 

What most people do not know is that Undelete also provides features to keep your data secure and visibility into who deletes files from file servers.

When a file is deleted, many assume that file data is now safe from being seen by others. Not so fast. When data gets deleted on a Windows volume, the data does not get removed. The space where that file data was residing is just marked as available for use, but the original file data is still there and will remain there until that space is overwritten by some other file data. That may or may not happen for quite a while. This means, that ‘deleted’ file data could still be potentially read. 

So, what do you do if you really want your file data gone when you delete it?  Undelete has the answer with two features. The first feature is “SecureDelete.”  When a file is deleted, SecureDelete will first over-write the file to help ensure it is unrecoverable. In fact, this is done by overwriting it with a specific bit pattern specified for this purpose by the U.S. National Security Agency (NSA) for the Department of Defense (DOD).  The second feature is “Wipe Free Space”, which will overwrite any free space on a selected volume, using the same specific bit patterns as SecureDelete to clear out any previously written data in that free space.

Now, with these two features, when you delete a file, you know it is now virtually impossible to read/recover any of that data from that volume.

Along with file security, there are customers using Undelete as another precautionary security: check how many files are being deleted from file shares and by whom. If they ever detect an abnormal, substantially high number of files being deleted, that raises a flag for them to investigate further.

Although Undelete is usually purchased to recover files, others use it to securely delete files and track back any deleted files to the person who did it.

Tags:

Data Protection | File Protection | Undelete

MEDITECH Hospital Speeds EHR & MS-SQL with V-locity® I/O Reduction Software

by Brian Morin 28. August 2017 10:06

Community Medical Center (CMC) had one initial requirement – find a FAL remediation solution for their MEDITECH electronic health record (EHR) application to maintain 24/7 availability and avoid downtime. What surprised them the most was the extent of the performance boost from using V-locity I/O reduction software.

“Our doctors and clinicians were losing too much time on basic tasks like waiting on medical images to load, or scanning images, or even just navigating from screen to screen within the application. The easy answer is to buy new server and storage hardware; however, that’s also a very expensive answer. When you’re a small hospital, you need to squeeze every last drop of performance out of your existing infrastructure. Since we don’t have the budget luxury of doing hardware refreshes every three years, we need to get at least five years or more from our storage backend,” said Joe Buckminster, IT Director, Community Medical Center.

Buckminster continued, “We initially purchased V-locity I/O reduction software to meet an availability requirement, but what surprised us the most was how much value it to added to our aging storage infrastructure by offloading a significant amount of I/O traffic. Not only did we get an immediate performance boost for MEDITECH, but we soon realized that we needed to try V-locity on our other Tier-1 applications like NextGen, MS-SQL, MS Exchange, Citrix XenApp, and others.”

Joe identified 35 key virtual servers that ran an assortment of different applications, like NextGen EHR (supported by a MS-SQL database), MS Exchange, Citrix XenApp, GE Centricity Perinatal, and others. In aggregate, V-locity offloaded 43% of all read traffic from storage and 29% of write traffic. With well over half a billion I/Os eliminated from going to storage, the median latency savings meant an aggregate of 157 days of cumulative storage I/O time saved across all the servers over a three-month period. When examining the last 24 hours from CMC’s single heaviest workload on a MS-SQL server, V-locity offloaded 48,272,115 I/O operations from storage (48% of read traffic / 47% of write traffic) – a savings of seven hours in storage I/O time

“There’s no way we would have achieved a 5-year lifecycle on our storage system without V-locity offloading so much I/O traffic from that subsystem. We had no idea how many I/O operations from virtual server to storage were essentially wasted activity due to Windows write inefficiencies chewing up IOPS or hot data that is more effectively served from available DRAM,” said Buckminster.

 

To read the full story on how V-locity I/O reduction software boosted their EHR and MS-SQL performance, read here: http://learn.condusiv.com/rs/246-QKS-770/images/CS-Community-Medical.pdf

Tags:

MEDITECH | V-Locity

Diskeeper Administrator Quick Start Guide

by Tim Warner, Microsoft MVP in Cloud and Datacenter Management 4. August 2017 06:31

Here's the situation: Sluggish application performance on one or more of your Windows servers had become so bad as to require intervention. You purchased a license for Condusiv® Diskeeper Server and observed such an amazing performance improvement that you want to deploy Diskeeper® on all your physical servers.

"But how can I centrally manage the application?" you wonder. And then you see that Condusiv Makes Diskeeper Administrator. Bingo!

Diskeeper Administrator gives you centralized control over all your managed servers. The solution enables you to deploy, configure and manage Diskeeper Server, Diskeeper Professional for desktop workstations, and some very early versions of V-locity for virtual machines (VMs). Condusiv plans to integrate SSDkeeper into Diskeeper Administrator sometime in the future.
Note: (V-locity and V-locity Management Console should be used for virtual servers)

For now, though, let's learn how to deploy and configure Diskeeper Administrator.

Deployment Prerequisites

In addition to purchasing a Diskeeper Administrator license, you should consider a volume-license purchase of Diskeeper Server to save money.
Diskeeper Administrator runs as a Windows service and is a client/server application that uses Microsoft SQL Server for back-end data storage. For a smooth installation experience, I suggest, but not required, having a SQL Server database engine instance already available on the network. Diskeeper Administrator supports the following database versions:

   - SQL Server 2005
   - SQL Server 2008
   - SQL Server 2012

You can use the free Express Edition if you want; in fact, the Diskeeper Administrator installer can automatically install SQL Server 2012 Express Edition. Notably, more recent SQL Server versions are not yet supported.

Like any client/server application, you need to keep firewall rules in mind. Diskeeper Administrator uses the following Transmission Control Protocol (TCP) ports:

   - 1434 (for SQL Server)
   - 31029
   - 31036
   - 30137
   - 31056
   - 31096
   - 31116
   - 31176
   - 31196
   - 31216

All your managed servers should have file and printer sharing enabled, which opens TCP ports 139 and 445, and exposes the ADMIN$ administrative share that is used for Diskeeper push installation. In addition, you should open ports 31038 or 31058 to facilitate management traffic. More details on port use is available in the DK Administrator’s online help.

Diskeeper Administrator Install and Setup

The Diskeeper Administrator installer is basically an InstallShield wizard "click-click-next" routine. The real work begins after you lay down the application binaries and start Diskeeper Administrator for the first time.

Speaking of Windows Server, you can install Diskeeper Administrator on any version from Windows Server 2008 R2 to Windows Server 2016, as well as Windows Client versions from Windows 7 to Windows 10. My environment runs Windows Server 2016 exclusively, and Diskeeper products all run just fine.

On first launch of the Diskeeper Administrator console (it's an honest-to-goodness Windows desktop application and not a browser portal), you'll see the following requirements dialog:

Continue reading the full Diskeeper Administrator Quick Start Guide »

 

RecentComments

Comment RSS

Month List

Calendar

<<  February 2019  >>
MoTuWeThFrSaSu
28293031123
45678910
11121314151617
18192021222324
25262728123
45678910

View posts in large calendar