Skip to main content

SAN these are top 20 benefits you need to know

In my previous post I have covered about fundamentals of SAN(storage area networks). The below are the list of top 20 benefits of storage area networks.

Greater performance: Current Fibre Channel SANs allow connection to disks at hundreds of megabytes per second; the near future will see speeds in multiple gigabytes to terabytes per second.

Increased disk utilization: SANs enable more than one server to access the same physical disk, which lets you allocate the free space on those disks more effectively.
Higher availability to storage by use of multiple access paths: A SAN allows for multiple physical connections to disks from a single or multiple servers.

Deferred disk procurement: That’s business-speak for not having to buy disks as often as you used to before getting a SAN. Because you can use disk space more effectively, no space goes to waste.

Reduced data center rack/floor space: Because you don’t need to buy big servers with room for lots of disks, you can buy fewer, smaller servers — an arrangement that takes up less room.

New disaster-recovery capabilities: This is a major benefit. SAN devices can mirror the data on the disks to another location. This thorough backup capability can make your data safe if a disaster occurs.

Online recovery: By using online mirrors of your data in a SAN device, or new continuous data protection solutions, you can instantly recover your data if it becomes lost, damaged, or corrupted.

Better staff utilization: SANs enable fewer people to manage much more data.

Reduction of management costs as a percentage of storage costs: Because you need fewer people, your management costs go down.

Improved overall availability: This is another big one. SAN storage is much more reliable than internal, server-based disk storage. Things break a lot less often.

Reduction of servers: You won’t need as many file servers with a SAN. And because SANs are so fast, even your existing servers run faster when connected to the SAN. You get more out of your current servers and don’t need to buy new ones as often.

Improved network performance and fewer network upgrades: You can back up all your data over the SAN (which is dedicated to that purpose) rather than over the LAN (which has other duties). Since you use less bandwidth on the LAN, you can get more out of it.

Increased input/output (I/O) performance and bulk data movement: Yup, SANs are fast. They move data much faster than do internal drives or devices attached to the LAN. In high-performance computing environments, for example, IB (Infiniband) storage-network technology can move a single data stream at multiple gigabytes per second.

Reduced/eliminated backup windows: A backup window is the time it takes to back up all your data. When you do your backups over the SAN instead of over the LAN, you can do them at any time, day or night. If you use CDP (Continuous Data Protection) solutions over the SAN, you
can pretty much eliminate backup as a separate process (it just happens all the time).

Protected critical data: SAN storage devices use advanced technology to ensure that your critical data remains safe and available.

Nondisruptive scalability: Sounds impressive, doesn’t it? It means you can add storage to a storage network at any time without affecting the devices currently using the network.

Easier development and testing of applications: By using SAN-based mirror copies of production data, you can easily use actual production data to test new applications while the original application stays online.

Support for server clusters: Server clustering is a method of making two individual servers look like one and guard each other’s back. If one of them has a heart attack, the other one takes over automatically to keep the applications running. Clusters require access to a shared disk drive;
a SAN makes this possible.

Storage on demand: Because SAN disks are available to any server in the storage network, free storage space can be allocated on demand to any server that needs it, any time. Storage virtualization can simplify storage provisioning across storage arrays from multiple vendors.


Popular posts from this blog

Blue Prism complete tutorials download now

Blue prism is an automation tool useful to execute repetitive tasks without human effort. To learn this tool you need the right material. Provided below quick reference materials to understand detailed elements, architecture and creating new bots. Useful if you are a new learner and trying to enter into automation career.
The number one and most popular tool in automation is a Blue prism. In this post, I have given references for popular materials and resources so that you can use for your interviews.
Why You Need to Learn RPA blue prism tutorial popular resources I have given in this post. You can download quickly. Learning Blue Prism is a really good option if you are a learner of Robotic process automation.
RPA Advantages The RPA is also called "Robotic Process Automation"- Real advantages are you can automate any business process and you can complete the customer requests in less time.

The Books Available on Blue Prism 
Blue Prism resourcesDavid chappal PDF bookBlue Prism…

Topologies in Apache Storm the concept you need to know

There are two main reasons why Apache Storm is so popular. The number one is it can connect to many sources. The number two is scalable. The other advantage is fault tolerant. That means, guaranteed data processing.
The map-reduce jobs process the data analytics in Hadoop. The topology in Storm is the real data processor. The co-ordination between Nimbus and Supervisor carried by Zookeeper What are topologiesThe jobs in Hadoop are similar to topology. The jobs run as per schedule defined.In Storm, the topology runs forever.A topology consists of many worker processes spread across many machines. A topology is a pre-defined design to get end product using your data.A topology comprises of 2 parts. These are Spout and bolts.The Spout is a funnel for topology Two nodes in StormMaster Node: similar to Hadoop job tracker. It runs on a daemon called Nimbus.Worker Node: It runs on a daemon called Supervisor. The Supervisor listens to the work assigned to each machine.Master NodeNimbus is re…

Testing in DevOps to maximize Quality

Testing is the critical phase in DevOps. The process of DevOps is to speed up the deployment process. That means there are no shortcuts in testing. Covering most relevant test cases is the main thing the tester has to focus.
Code Quality RequirementsGood maintainable codeExhaustive coverage of casesTraining documents to Operations teamFewer bugs in the bug trackerLess complex and no redundant code Testing Activities in DevOpsThe team to use Tools to check the quality of codeStyle checker helps to correct code styleGood design avoids bugs in productionCode performance depends on the code-qualityBugs in production say poor testing  Tester Roles in DevOpsGood quality means zero bugs in production.Design requirements a base to validate testing results.Automated test scripts give quick feedback on the quality of code. Right test cases cover all the functional changes. The Bottom LineThe DevOps approach is seamless integration between Development and Operations without compromising th…