Scaling Storage for Databases — How Did We Get Here?
We’ve seen the impact of digitization and the proliferation of data change just about everything in our world. And, as digitization grew, so did the demands on our infrastructure. What once were ‘simple’ computing devices have now become complex networks of computers to calculate weather patternsmodel genetic researchand map the spread of viral outbreaks. But, it’s only the beginning. Yet, how did we get here? What did this evolution look like and where are we today? Here are a few thoughts on how we got here and how organizations can successfully architect storage for scaling databases.
Years in the Making
At first, we started with single CPU systems like the HP 9000 I70s (yes, there were others before this). These became the K-Class Servers. Then, the Sun rose and we had the E10K, the V-Class, and the P-Series. While we kept scaling-up CPU resources, and to a certain degree memory resources, it simply wasn’t enough. To get bigger and faster, we started scaling out.
Now, clusters of CPUs and memory were linked together with what felt like gauze, thirty-weight oil, and ball-bearings, and the results were impressive. The speed of these highly specialized systems helped to deliver new, deeper insights, that drove better decision making for science, health, and businesses.
Yet, the world needed more. As a next step, investigating how to get more performance, the industry saw that Moore’s Law had been diligent with CPUs and memory, but had not had the same results with the storage and retrieval platforms. And, as a result, this is where we started to scale-up and scale-out our storage.
It is hard to believe, but at first we were only scaling from megabytes to 100s of megabytes. In the 90s, gigabytes were mind boggling, and then things took off. Today, we have multi-petabyte systems available off the shelf and massive object storage to land-lock seas of data. We have companies building exabyte scale solutions, and we are already creating the building blocks to fully enable the zettabyte age.
Scaling Databases – Scale-Up or Scale-Out?
Now, scaling is all about the data. To harness insights, you need more data. So, companies are looking at keeping all of their data and getting it as close to the CPUs as possible for better performance.
While everything is scaling – and some products requiring you to choose whether to scale-up or scale out, what do you do with databases and how do you decide which is right for you?
Scaling out provides access to additional storage, but increases the overall performance – both latency and bandwidth; whereas, scaling-up each of these units provides petabyte capacities to handle your most challenging growth models.
We believe you should not be limited. You need to be able to do both when it suits your business objectives and goals, along with you organization’s best practices for data strategy.
For example, our hybrid, all-flash and NVMe IntelliFlash arrays provide the option for both architectures; however, there are some advantages to one model over the other that you should consider:
• Redundant Controllers
• Expand by adding controllers and storage
• Performance increases vs. performance could improve
• Complexity and power increases
Again, the decision should be based on your business and data strategy goals.
Oracle® Grid Infrastructure (ASM, Automatic Storage Manager) – Example Use Case
Let’s look at a real world business application example. If you are, or you plan, to leverage Oracle Grid Infrastructure (ASM, Automatic Storage Manager), this application allows you flexibility when leveraging IntelliFlash.
For example, if your goal is to achieve as much performance as possible, looking at the chart above, choosing a scale-out architecture would be the most beneficial. At near 25GB/s and 200µs of latency, you can scale the N-Series system up to 140GB/s of bandwidth with 1/2 rack solution.
How IntelliFlash Supports Scale-Out and Scale-Up
With the increase in initiatives from Oracle AI / ML and Microsoft AI/ML solutions, IntelliFlash provides the flexibility to scale – whether out or up – to meet the demands of these products.
From an operating system that simplifies data management to data services (i.e. replication, snapshot, live LUN migration, and cloud integration), organizations are easily able to scale your application and data needs.
Whether you need to scale-out for flat latency and/or you need huge bandwidth to process data as fast as your edge devices can create it, with the N-Series arrays from Western Digital, you have the freedom to choose the infrastructure that best fits your needs, when you need it
Simply putting Data at the Center of your organization helping you deliver Speed. Insight. Decisions.
Ready to Tip the (Data) Scale?
Today’s data needs demand extreme performance at low latency. Through SAS, NVMe, and hybrid Intelliflash flash arrays, our solutions help enterprises balance performance with cost to scale-up and scale-out.
• We tested real-world Oracle performance on NVMe. Here are our test results
• Want to learn more about how to manage your data growth with Oracle or SQL Server®? Check out a recent webinar where we shared how IntelliFlash NVMe simplifies data management challenges. We jumped into the traditional infrastructure elements, such as capacity, performance, and consumption of data, and how Oracle and SQL Server databases have developed scale-out and scale-up approaches.
• Intelliflash combines flash, NVMe, and advanced data services. Watch the video
The post Scaling Storage for Databases — How Did We Get Here? appeared first on Western Digital Corporate Blog.
The Human Side of the Fourth Industrial Revolution
A few small sensors and a Raspberry Pi control the watering system in Taweesak Phuengprasit’s family’s farm in the Ang…
Innovation for Independence
Smart home data and automation assist those with special needs Home automation technology is making life easier and more convenient….
The Disks on the Bus Go Round and Round: How AngelTrax Is Re-envisioning Safety
Anna England and her two daughters each leave the house at different hours of the morning, but they are all…
Drive the Future
Data storage helps the automotive industry take the next step forward Car manufactures are building smarter technologies into autonomous vehicles…
The Race to Seal Helium HDDs
Innovation is rarely quick and often only obvious in hindsight. Helium has been one of the greatest breakthroughs for high-capacity…
A Taste of Data: How Data Is Revolutionizing the Alcohol Industry
Since antiquity, humans have made alcohol. It’s part of our ancestry and our every day, and there’s always been a…
Five Reasons the 2019 RISC-V Summit is a Can’t-Miss Event
Whether you’re a veteran RISC-V enthusiast or brand new to the world of open-source instruction set architecture (ISA), the RISC-V…
The IoT Evolution – Top 9 IoT Use Cases of 2019
As smart sensors are placed in billions of connected devices around the world, new IoT applications are generating massive streams…
Driving to Data-Centric Architectures and 1B RISC-V Cores
Two years ago we kicked off our commitment to open-source innovation by announcing our goal to transition over one billion…
In 2039, Could Fully Autonomous and Connected Cars Exist?
Recently, I gave the opening remarks and sat in on a panel at A Data-Driven Futurean automotive industry event hosted…
2019 Data Center Year in Review
The era of digital business transformation has been in full thrust in 2019, as emerging applications and data center infrastructure…
Industrial-Grade Storage Enables Drones for Search and Rescue Teams
First things, first: I’m a proud member of the El Dorado County Search and Rescue (EDSAR) organization and have been…