SIOS SANless clusters

SIOS SANless clusters High-availability Machine Learning monitoring

  • Home
  • Products
    • SIOS DataKeeper for Windows
    • SIOS Protection Suite for Linux
  • News and Events
  • Clustering Simplified
  • Success Stories
  • Contact Us
  • English
  • 中文 (中国)
  • 中文 (台灣)
  • 한국어
  • Bahasa Indonesia
  • ไทย

Reframing Early Computer Science Education: The Soft Skills of Solution Design Part 1

November 3, 2025 by Jason Aw Leave a Comment

Reframing Early Computer Science Education The Soft Skills of Solution Design

Reframing Early Computer Science Education: The Soft Skills of Solution Design Part 1

The Relationship of Rhetoric and Technical Design

When I was in college, my first year in the school of computer science had me in multiple courses directed towards writing code and completing lab exercises. I would classify this first class as a “coding class”.

My first year studying computer science was all about writing code, countless lines written for common beginner projects, and an intense focus on syntax and the core features of C as a programming language. Lab project after lab project, I wrote more code in that year than I had at any time leading up to it. The first year’s curriculum was just a year of coding. It wasn’t until my second year in the school of computer science that I took my first true course in software engineering. The distinction wasn’t about the course title – I think of that second-year class as my first real computer science course because it introduced a core concept of software engineering: solution design.

Rhetoric: The Hidden Foundation of Software Design

Though I did not realize it at the time, I had already been exposed to the principles of software engineering earlier in my education. These principles were introduced through a course on the rhetorical analysis of literature. The class revealed that communication relies on rhetorical patterns, the structures that carry a message toward understanding and purpose. Understanding why a written work is effective requires understanding the rhetorical patterns used in its construction. Works that were effective in their goal incorporated rhetorical techniques and structural patterns that were complementary and cohesive with respect to the goal of the piece. The unsuccessful works jumped between techniques and used patterns that had the same goal but were incompatible with one another. Consequently, they read as dissonant and ineffectual – sometimes they were just confusing. Successful rhetoric is successful because it was intentional. The importance of each of the work’s structural elements was apparent due to the patterns present when the work was planned.  In turn, these features resulted in clear, effective literature.

Why Understanding Design Principles Enables Long-Term Success

Technical fields, at a high level, have a common goal of producing reliable and maintainable solutions. Successful projects from an engineer or IT professional parallel effective rhetoric. The production of an effective and nontrivial solution requires that the solution first go through a design phase. Design prioritizes the use of patterns that are cohesive in concept and purpose, which is the foundational step in creating a solution that is understandable.

Maintainability follows understanding; when personnel understand the design and how each component of the implementation relates to the design, they are empowered to perform upkeep that follows the patterns and principles of the design. Implied in all of this is the assumption that design documentation is readily available and kept updated so teams may form an understanding to guide their actions. Then, so long as the respect for the design is upheld, the solution can continually be maintained via the understanding brought forth due to intelligent design.

How Purposeful Design Prevents System Decay

Given an architecture outlined with clear design principles and cohesive design patterns, a solution can see maintainers come and go throughout the solution’s life and still fulfill its purpose. Conversely, it is the experience of many engineers tasked with maintaining a legacy solution that lacks documentation or a clear design to puzzle over, and potentially break, the solution. Effective solutions are effective because they were designed intentionally. Solutions are effective when the patterns present in the design communicate the pursuit of purpose. The design is the vehicle by which one understands how each element’s role achieves the solution’s purpose. When elements of the solution are implemented in conjunction with a design that is cohesive in concept and purpose, solutions can be relied upon throughout their lifetime of maintenance and the iterations of improvements to come.

Author: Philip Merry, CX – Software Engineer at SIOS Technology Corp.

Reproduced with permission from SIOS

Filed Under: Clustering Simplified Tagged With: software design

How to Cut SQL Server HA/DR Costs and Gain Advanced Features

October 21, 2025 by Jason Aw Leave a Comment

How to Cut SQL Server HADR Costs and Gain Advanced Features

How to Cut SQL Server HA/DR Costs and Gain Advanced Features

Microsoft SQL Server is vital for mission-critical applications, making high availability (HA) and disaster recovery (DR) essential. However, Enterprise Edition licensing and SAN-based clusters drive up costs and complexity. This white paper reveals how organizations can cut licensing costs, eliminate single points of failure, and unlock advanced flexibility, all without relying on Enterprise Edition or expensive SANs.

Reproduced with permission from SIOS

Filed Under: Clustering Simplified Tagged With: High Availability and DR, SIOS Datakeeper, SQL Server

Commonalities between Disaster Recovery (DR) and your spare tire

October 14, 2025 by Jason Aw Leave a Comment

Commonalities between Disaster Recovery (DR) and your spare tire

Commonalities between Disaster Recovery (DR) and your spare tire

In our recent blogs, we’ve drawn some interesting parallels between cars and DataKeeper. These posts have explored topics such as:

  • Transitioning from LifeKeeper to Windows Server Failover Clustering (or vice versa)
  •  Maximizing the efficiency of your ‘GET’ commands in DataKeeper
  • Comparing your car dashboard to the DataKeeper User Interface (UI)

Let’s keep that theme rolling (pun intended)

Understanding the Role of a Spare Tire (and a DR Node)

Let’s give a brief intro on the function of a spare tire and the function of a DR node in a DataKeeper clustered environment running Windows Server Failover Clustering™.

A spare… will temporarily replace a damaged tire, allowing you to reach a repair shop, home, or other destination, saving you time and avoiding being towed ($$$) or stranded. Though convenient, temporary spares have limits on longevity and speed.

Understanding the Role of a Disaster Recovery Node

A Disaster Recovery node . . .  is typically a standby node (spare) that contains applications and data, often located in a different region from its primary location to protect against outages/disasters, man-made or natural.

There are endless pros and cons for both.  I’ve named just a few for the sake of readership . . .

Drawing Parallels Between Your Spare Tire and a DR Node

Pros (with a spare) Cons (without a spare)
Reduce being stranded Delays, stranded overnight
Avoid Roadside Assistance Roadside service may take hours
Mobile again to go get it fixed permanently Must wait for a tow or other means to get the repair done, which can be costly
Pros (with DataKeeper) Cons (without DataKeeper)
Streamline failover without manual intervention Need to rebuild systems, restore data manually
Reduce risk of data loss SLAs not met, loss of sales, penalties
Maintaining customer trust Not meeting customer expectations reduce confidence

In this blog, we can draw a clever analogy between Disaster Recovery (DR) in DataKeeper clustered environments and the humble “doughnut” tire in your car.

Both serve as critical safety nets in moments of crisis, ensuring you can recover quickly and avoid prolonged downtime.

Why a Reliable DR Solution Matters More Than Ever

Just as a spare tire ensures you can keep driving after a flat, a DR node provides critical backup infrastructure to keep your business running smoothly in the face of outages, cyberattacks, or natural disasters.

In today’s fast-paced digital world, downtime can result in lost revenue, damaged reputation, and even legal liabilities—making the need for a reliable DR solution more crucial than ever.

A DR node acts as a safety net, allowing businesses to recover quickly and minimize disruptions to operations. For customers, investing in a DR node is not just about mitigating risk; it’s about ensuring peace of mind, protecting valuable data, and maintaining trust with clients and stakeholders.

Keep Your Business Rolling with DataKeeper

In short, a Disaster Recovery node is the cornerstone of resilience, empowering businesses to stay agile and focused no matter what challenges arise. Whether it’s a spare tire or a Disaster Recovery node, preparedness is the key to staying on track when life throws unexpected challenges your way. Just like you wouldn’t drive without a spare, don’t run your business without a DR plan. Request a demo to see how DataKeeper keeps your operations moving.

Author: Greg Tucker Senior Product Support Engineer at SIOS Technology

Reproduced with permission by SIOS

Filed Under: Clustering Simplified Tagged With: DataKeeper, disaster recovery

Unlocking Near-Zero Downtime Patch Management with High Availability Clustering

October 3, 2025 by Jason Aw Leave a Comment

Unlocking Near-Zero Downtime Patch Management with High Availability Clustering

Unlocking Near-Zero Downtime Patch Management with High Availability Clustering

Patch management is one of the toughest balancing acts in IT. Every month or quarter, OS and application vendors release updates with critical security fixes. These patches need to be tested and applied quickly — but rushing the process risks instability, and delaying it increases vulnerability. For organizations running mission-critical applications, the stakes are even higher.

That’s why IT leaders are increasingly turning to high availability (HA) clustering to streamline patch testing and deployment, while keeping downtime to a minimum.

Why Patch Management Is So Challenging

  • Testing takes time and resources. QA environments aren’t always available, and teams may feel pressure to shortcut testing just to keep up.
  • Cyberattacks move fast. Zero-day exploits are weaponized within hours of a patch release. According to the Ponemon Institute, 57% of breaches are attributed to unpatched vulnerabilities.
  • Downtime is costly. Whether planned or unplanned, downtime averages $5,600 per minute (Gartner). In industries such as healthcare, aviation, and manufacturing, even a brief outage can have significant financial and safety implications.

The challenge is clear: organizations must patch faster, test thoroughly, and minimize disruptions.

How HA Clustering Transforms Patch Management

High availability clustering pairs a primary server node with a secondary node. Advanced clustering software continuously monitors the environment — applications, OS, storage, and networks. If a failure occurs, operations seamlessly move to the secondary node without downtime.

This same architecture enables a “rolling upgrade” approach for patching:

  1. Patch the secondary node while the primary node continues to run.
  2. Test the update on the secondary node before making the switch.
  3. Fail back if needed — if issues are found, operations instantly continue on the primary node.
  4. Cut over if successful — if tests pass, operations shift to the secondary node, and the primary can be patched next.

The result: organizations can apply updates faster, avoid risky shortcuts, and keep systems available 24/7.

Strengthening Security, Compliance, and IT Resilience with HA Clustering

Modern regulations, such as HIPAA, PCI DSS 4.0, and NIST 800-53, require timely patching. At the same time, high-profile incidents (such as the CrowdStrike update failure) have shown the danger of rushed, untested updates.

By integrating HA clustering into patch management strategies, IT teams can:

  • Meet compliance requirements without sacrificing uptime.
  • Reduce risk from patch-related failures.
  • Strengthen overall IT resilience against cyberthreats.

Near-Zero Downtime Patch Management for Mission-Critical Applications

The old trade-off between speed and stability in patching no longer exists. With high availability clustering, IT teams can patch quickly, test safely, and keep mission-critical applications online, all while reducing downtime to near zero.

If your organization struggles with patch management, HA clustering may be the key to safer updates and stronger resilience.

Ready to eliminate downtime from your patching process? Request a demo of SIOS High Availability Clustering and see how your team can patch faster, stay compliant, and keep critical applications running 24/7.

Author: Ben Roy, Marketing Specialist at SIOS

Reproduced with permission from SIOS

Filed Under: Clustering Simplified Tagged With: High Availability, Patch Management

How to Safely Combine DataKeeper for Linux with Backup and Replication Tools

September 26, 2025 by Jason Aw Leave a Comment

How to Safely Combine DataKeeper for Linux with Backup and Replication Tools

How to Safely Combine DataKeeper for Linux with Backup and Replication Tools

When using other Backup or Replication Software with DataKeeper for Linux, the purpose of DataKeeper is to replicate data between servers in a cluster, ensuring all relevant servers have the most up-to-date copy of data. This is crucial when a server experiences unplanned downtime, and LifeKeeper is able to ensure critical applications are highly available and can maintain uptime with the use of DataKeeper.

When combining DataKeeper with other backup or replication software, it’s essential to confirm compatibility to avoid conflicts. Replication software can interfere with DataKeeper’s resynchronization, sometimes due to the order in which replication processes begin. While aiming for maximum uptime and availability is beneficial, it’s critical to verify that such measures will maintain your cluster in an optimal state.

How to Test DataKeeper for Linux with Backup and Replication Software

It’s important to test the compatibility of the replication software being used alongside DataKeeper to ensure its functionality. Below is a list of items you can check to verify functionality.

1. Test on a QA cluster.

Before using both backup/replication software on your production cluster, create a QA cluster environment with DataKeeper to run tests on.

A QA cluster is beneficial for running tests before introducing anything new into your production cluster. This helps with avoiding issues that would arise on your production cluster by being proactive with catching and/or fixing any issues that arise on your QA cluster.

2. Complete basic functionality test.

A couple of basic tests should be completed with DataKeeper as the only replication software installed. This is a sanity check before verifying continuing with any other software.

Base tests should include testing for a successful switchover and failover. Visit the link below for steps to confirm switchover can be successfully performed.

https://docs.us.sios.com/spslinux/9.9.1/en/topic/testing-your-datakeeper-resource-hierarchy

3. Complete basic functionality tests with other software.

Run the same tests mentioned above while the software is backing up/replicating your data, and after the software has completed backing up/replicating your data.

To be able to use the software with DataKeeper, it’s important that all these functionality tests pass.

Using GenApp Resources to Manage Backup and Replication Processes with DataKeeper for Linux

If testing yields unsuccessful results, it is possible to create a Generic Application (GenApp) to start and stop the relevant processes during a switchover

  • A GenApp can be used in the hierarchy to restore and remove the process used by the replication software to handle the order in which the software runs.
    • A hierarchy determines the relationship between resources. Top-level resources depend on bottom-level resources to create a dependency relationship. When a hierarchy is taken out of service, LifeKeeper takes a top-down approach, removing the top-level resources before the bottom-level resources. When a restore is issued, LifeKeeper takes a bottom-up approach to restore the bottom-level resources before restoring the top-level resources.

With this understanding, two GenApps would be created, one as a top-level resource and the other as a bottom-level resource. This configuration ensures that when the hierarchy comes into service, the bottom-level GenApp will stop the process, and the top-level GenApp will start it. When the hierarchy is being removed, the only action would be for the bottom-level resource to stop the process.

  • Read more about creating a GenApp in the link below.

https://docs.us.sios.com/spslinux/9.9.1/en/topic/creating-a-generic-application-resource-hierarchy

Ensuring DataKeeper Cluster Compatibility and Preventing Downtime

Ultimately, testing and verification are key before introducing more backup or replication software into your DataKeeper Cluster. These steps are intended to avoid downtime by providing a list of items to complete to make sure your configuration is in order before being introduced into your production environment. Before integrating additional backup or replication software with your Linux DataKeeper Cluster, thorough testing and verification are essential. Completing these steps ensures your configuration is properly set up and helps prevent downtime when introduced into your production environment.

Ready to see how SIOS can help you simplify high availability and ensure seamless backup and replication with DataKeeper for Linux? Request a demo today.

Author: Alexus Gore, Customer Experience Software Engineer

Reproduced with permission from SIOS

Filed Under: Clustering Simplified Tagged With: backup, replication

  • « Previous Page
  • 1
  • 2
  • 3
  • 4
  • …
  • 110
  • Next Page »

Recent Posts

  • 99.99% Uptime: Balancing High Availability and Maintenance
  • Video: EGGER Achieves 99.99% Uptime with SIOS LifeKeeper for Linux
  • Resilient Together: How Partnerships Drive Modern Disaster Recovery
  • Three Keys to Mastering High Availability in Your On-Prem Data Center
  • Why High Availability Matters in Manufacturing 4.0

Most Popular Posts

Maximise replication performance for Linux Clustering with Fusion-io
Failover Clustering with VMware High Availability
create A 2-Node MySQL Cluster Without Shared Storage
create A 2-Node MySQL Cluster Without Shared Storage
SAP for High Availability Solutions For Linux
Bandwidth To Support Real-Time Replication
The Availability Equation – High Availability Solutions.jpg
Choosing Platforms To Replicate Data - Host-Based Or Storage-Based?
Guide To Connect To An iSCSI Target Using Open-iSCSI Initiator Software
Best Practices to Eliminate SPoF In Cluster Architecture
Step-By-Step How To Configure A Linux Failover Cluster In Microsoft Azure IaaS Without Shared Storage azure sanless
Take Action Before SQL Server 20082008 R2 Support Expires
How To Cluster MaxDB On Windows In The Cloud

Join Our Mailing List

Copyright © 2025 · Enterprise Pro Theme on Genesis Framework · WordPress · Log in