|
|
|||||||
|
Welcome to the V I P E R L A I R .com - Forums. You are currently viewing our boards as a guest which gives you limited access to view most discussions and access our other features. By joining our free community you will have access to post topics, communicate privately with other members (PM), respond to polls, upload content and access many other special features. Registration is fast, simple and absolutely free so please, join our community today! If you have any problems with the registration process or your account login, please contact contact us. |
|
|
|
Thread Tools | Display Modes |
|
#1
|
|||
|
|||
|
Data Corruption 101
Corruption 101
There sometimes seems to be a few hundred ways to munge your data my personal list of probability based on The Risks To Your Data @ the PC Guide (with a few additions, and split into hardware, software and filesystem) Hardware Failure Memory Errors: With so many systems today running without error detection or correction on their system memory, there is a chance of a memory error corrupting the data on the hard disk. It is rare for it to happen, but it does happen. Test your RAM with memtest86 and or Memtest, consider a board that supports ECC RAM[/url] Power Loss: Losing power at the wrong time, such as when you are doing sensitive work on your hard disk, can easily result in the loss of many files Use a high quality PSU with stable voltage, employ a UPS or other line conditioning, avoid hard restarts Cables not included on the PCGuide's original list it is none the less an importatnt possibility, especially with todays transfer speeds, there is a very real reason that the industry is adopting SATA over PATA see post below for a full discussion and links System Timing Problems: Setting the timing for memory or cache access too aggressively, or using a hard disk interface transfer mode that is too fast for the system or device, can cause data loss. This is often not something that will generally be realized until after some amount of damage has been done. see tRAS below, bump back an overclock, try a different divider, dont overclock if you cant lock the PCI bus, I would bump the probability of this up a few slots if the machine is overclocked, otherwise I ve put it here Resource Conflicts: Conflicts resulting from peripherals that try to use the same interrupt requests, DMA channels or I/O addresses, can cause data to become corrupted. Review PIRQ routing and manual assignment of IRQs in the worse case senerio, chipset specific but a good overview The Hard DriveTest with the manufacturers Diagnostic, most of the rest is preventative, proper handling vibration free and cool environment, with clean air and stable floor Software Failure Busmastering Drivers Filesytem Corruption Power Issues There are three basic areas of power problems 1.Source Power Brown outs, blackouts, spikes\surges ect. see > Power Conditioning and DIY UPS @ Dans Data, for the basics In this category I would also place power issues due to pilot error, hard restarts and shorts, avoid both. Shutdown properly and pay attention when mounting your motherboard and routing power cables. 2. Under Power: Basically too many components for the power supply, dont be decieved by wattage figures, its the amount of amps per rail that is really important. See > Choosing the right Power Supply takaman's Power Supply Calculator rev0.61x to determine the amps you need per rail 3. Voltage Stability Pretty much the all the following [H]ardcore PSU info (Charts) http://terasan.okiraku-pc.net/dengen/tester/index.html http://terasan.okiraku-pc.net/dengen/tester2/index.html In Japanese :p But the graphs speak volumes and the PSU are identified in English the "translation" links are below, but they really dont add much and take considerable time to load http://www.worldlingo.com/wl/transla...r%2Findex.html http://www.worldlingo.com/wl/transla...2%2Findex.html Continuous Power vs. Peak Power at Spin-Up [quote]12V power profile (current vs. time) of an IDE/ATA hard disk at startup. You can see that the peak power draw is over quadruple the steady-state operating requirement. The graph appears "noisy" due to frequent oscillations in current requirements [/QUOTE Peak vs. Continuous Power
Quote:
http://terasan.okiraku-pc.net/dengen/tester/index.html the consistent voltage instability at startup and shortly thereafter in those graphs Winbond Launches New Bus Termination Regulator April 4th 2003 "Winbond Electronics Corporation, a leading supplier of semiconductor solutions, today launched the W83310S, a new DDR SDRAM bus termination regulator. The solution, new to Winbond's ACPI product family, is aimed at desktop PC and embedded system applications with DDR SDRAM requirements. Computer systems architectures continue to evolve and are becoming more complex; CPU and memory speeds continue to increase ever more rapidly with every technology turn. More and more high current/low voltage power sources are required for PC systems. This is particularly true for high-speed components such as CPU, memory, and system chipsets. The performance of these components is highly dependent upon stable power. Therefore, motherboard designers require accurate, stable, low-ripple and robust power solutions for these components. Many system designs use discrete components to implement bus termination functions. This approach creates several problems including poorer quality load regulation; higher voltage-ripple, increased usage of board space and inconsistent designs when different discrete components are used." and just to reinterate this point one more time http://www.anandtech.com/showdoc.html?i=1774&p=8 "the majority of damaged RAM returned to memory manufacturers is destoryed by fluctuations in the voltage." the transient response is the critical measure, unfortunately its not a metric that is commonly supplied with the PSU specs Transient Response: As shown in the diagram here, a switching power supply uses a closed feedback loop to allow measurements of the output of the supply to control the way the supply is operating. This is analogous to how a thermometer and thermostat work together to control the temperature of a house. As mentioned in the description of load regulation above, the output voltage of a signal varies as the load on it varies. In particular, when the load is drastically changed--either increased or decreased a great deal, suddenly--the voltage level may shift drastically. Such a sudden change is called a transient. If one of the voltages is under heavy load from several demanding components and suddenly all but one stops drawing current, the voltage to the remaining current may temporarily surge. This is called a voltage overshoot. Transient response measures how quickly and effectively the power supply can adjust to these sudden changes. Here's an actual transient response specification that we can work together to decode: "+5V,+12V outputs return to within 5% in less than 1ms for 20% load change." What this means is the following: "for either the +5 V or +12 V outputs, if the output is at a certain level (call it V1) and the current load on that signal either increases or decreases by up to 20%, the voltage on that output will return to a value within 5% of V1 within 1 millisecond". Obviously, faster responses closer to the original voltage are best."
__________________
Entia non sunt multiplicanda praeter necessitatem Dual 244s, K8W, Quadro FX3000, 4GB PC2700, SX6000, SANnet 3200 |
|
#2
|
|||
|
|||
|
Cables
there is definately a reason why SATA is being adopted, ATA\IDE\EIDE\ATAPI is an unterminated standard and as speeds increase that causes more and more problems, especially if cheap cables are employed, complex device configurations, hotswap\removable drivebay bridgecards and poor cable routing a few (:p) links and excerpts: Standard (40-Conductor) IDE/ATA Cables In many ways, the cable is the weak link in the IDE/ATA interface. It was originally designed for very slow hard disks that transferred less than 5 MB/s, not the high-speed devices of today. Flat ribbon cables have no insulation or protection from electromagnetic interference. Of course, these are reasons why the 80-conductor cable was developed for Ultra DMA. However, even with slower transfer modes there are limitations on how the cable can be used. The main issue is the length of the cable. The longer the cable, the more the chance of data corruption due to interference on the cable and uneven signal propagation, and therefore, it is often recommended that the cable be kept as short as possible. According to the ATA standards, the official maximum length is 18 inches, but if you suspect problems with your hard disk you may find that a shorter cable will eliminate them. Sometimes moving where the disks are physically installed in the system case will let you use a shorter cable Warning: There are companies that sell 24" and even 36" IDE cables. They are not recommended because they can lead to data corruption and other problems. Many people use these with success, but many people do a lot of things they shouldn't and get away with it. :^) Ultra DMA (80-Conductor) IDE/ATA Cables There are a lot of issues and problems associated with the original 40-conductor IDE cable, due to its very old and not very robust design. Unterminated flat ribbon cables have never been all that great in terms of signal quality and dealing with reflections from the end of the cable. The warts of the old design were tolerable while signaling speeds on the IDE/ATA interface were relatively low, but as the speed of the interface continued to increase, the limitations of the cable were finally too great to be ignored. that "upgrade happened at 66MB/s burst, we are now at the same speed as the PCI bus for burst rates 133MB/s Fancy IDE leads - The Terrible Truth The spec mandates such short cables for two reasons. Reason one - practically all IDE cables are unshielded. There's nothing around the conductors but insulation. Electromagnetic radiation goes straight through insulation. So external interference from the rest of your computer's giblets can influence the signal on your IDE leads. Unshielded cables act like antennas. Generally speaking, the longer you make 'em, the more energy they can pick up from their environment. Reason two - IDE cables are unterminated. "Termination", in the electrical sense, is essential to provide "impedance matching", which in English is what you have to do to stop the signal from reflecting off the end of the cable like a wave that hits the end of a bathtub. Electric current does not move instantaneously down a wire. It travels at nearly the speed of light, but when you've got thirty-three and a third million clock pulses per second - which is the speed of the IDE bus - even light in a vacuum only moves a hair under nine metres per clock pulse. So if you're fooling around with, say, a double-the-rated-length 900mm IDE lead, there's an end-to-end signal delay in it of about a tenth of a clock pulse. The signals you want your drives and your motherboard to be able to hear will be significantly blurred by delayed reflections from each end of the cable. Transfer your data at twice or three times the UDMA/33 speed - as UDMA/66 and 100 do - and reflected signals get more and more out of step with the real signal, and do it more and more harm. Serial ATA and the 7 Deadly Sins of Parallel ATA Critical Limiting Factors in Parallel Design There are some fundamental differences between serial and parallel buses, more importantly, there are some critical limiting factors in the design and implementation of any parallel bus. 1. Non-Interlocked (source synchronous) clocking 2. 3.3 V high-low signaling with 5V legacy tolerance 3. Cabling constraints 4. Connector legacy 5. Termination 6. Command queuing 7. PCB Design 3. Cable Design Issues: Cross-Talk and Ground Bouncing vs.Ringing Each signal propagating through a data line makes the data line act like the inductor of a transformer. That is, each voltage swing generates a dynamic electromagnetic field, that, depending on cable length and proximity will induce another signal in adjacent data lines. This cross-talk adds noise to data lines and can produce errors by generating false positives or negatives simply by induction of voltage swings in data lines. Another problem with parallel pathways is the phenomenon of simultaneously switching outputs (SSO) noise. As we explained in detail in our reviews of the i845 and the SIS645 chipsets, SSO noise becomes really problematic if the majority of signals switch from high to low since this can induce ground bouncing. On the chipset level, workaround in form of dynamic bus inversion (DBI) is feasible, that is, instead of switching all bits, only the reference bit is switched simultaneously at the sender and receiver end which has the same net effect, namely, that the system does not see the reference switch but thinks that all other lines have switched. DBI, however requires an additional latency cycle and this is where the 40 ns clock cycle time starts to look really ugly. ATA not so Frequently Asked Questions Or: Why Ribbon Cables are unsuitable for RF transmission of data The following article was written by snn47 to address some of the issues associated with standard ribbon cables and the use of e.g. removable drive racks as an attempt to share some insight into factors that can adversely affect the life or reliability of of desktop Hard Disk Drives. Specifically, issues like why some drives are working in some systems and not in others, the impact of cable routing and why is it that the drive manufacturers always recommend using their own cables (if supplied with the drive). (emphasis mine) Any RF system has a limited tolerance for distortion of signals, which, in the worst case, can destroy some of the semiconductor components. While a certain amount of variation is part of any systems specification, one needs to remember that ATA was never intended to handle today's data rates. ATA or Advanced Technology Attachment started as the usual run of the mill or: "just a system at the lowest possible price point that will work most of the time without the need for huge financial investments". The problems started when the system was forced to handle higher and higher clock and data rates within the original design limitations. Keep in mind that the latest ATA-PI7 specifications allow data rates of 133 MB/sec, which is 44-times faster than the original ATA transfer of 3 MB/sec. This increase in speed makes it necessary to enforce minimum tolerances and detailed specifications to allow for the manufacturing of affordable systems with minimum compatibility problems. |
|
#3
|
|||
|
|||
|
these are just a few excerpts, I would highly advice that everyone give them a good read, there ARE good rounded ATA cables RD3XP Super Shielded
"RD3XP is made from ATA 100/133 High impedance flat cable cut into 8 layers of 10 cable wires, with a ground wire and signal wire alternatively, and folded in zigzag-piled so that each signal wire is surrounded by 4 ground wires." but like their SCSI counterparts, they aint cheap, there are also high quality flat cables (you buy a $300 RAID card, and they dont ship you crappy PVC cables, they are either Teflon or Thermoplastic Olefin (TPO) Up until a little while ago I would have said ant investment made in high quality cables was money well spent, however with the introduction of SATA, that doesnt necessarily hold true anymore unless your dealing with critical data (in which case you should be running ECC RAM) or your actually experiencing problems a further excerpt from ATA not so FAQs Preliminary Conclusions and Possible Cure Reasons for changes in the propagation impedance, cross-coupling between adjacent signal wires and signal-velocity from one setup to another are : Impedance of the drive and controller in high/low signal level will be different for different models. Reflection of signals that garble the pulse, due to incorrect termination impedance or impedance-inconsistencies from the controller to the drive meaning the Impedance from the controller and the drive(s) differ. If there is a a second drive (connector present/connected) the impedance will fluctuate at this point. A. Only one HDD per controller channel. B. Use a cable with only 2 connectors. -Signal delay will increase with the length of the flat-ribbon-cable propagation of the signals were intended for a max. flat-ribbon-cable length of 18" with ~ 5ns/m would be 2.3ns delay. C. shorten the cable whenever possible. D. If the case requires long cables consider mounting just the HDD closer to the connectors of the controller or consider exchanging the usual desktop case, for a 19" case. Mount the HDD just above or below the PCB-controller-connector to allow you to reduce the length of the flat-ribbon-cable to a few cm. Flat-ribbon-cable with different isolation material (higher/lower eR) and change in the conductor diameter will change the ratio of (2D/d). Are rounded cable used? E. Try exchanging the cable against another type/brand of flat-ribbon-cable. Is the flat-ribbon-cable at some point parallel to a conducting grounded surface? F. Try a different routing of your flat-ribbon-cable away from a ground-plane, Was the cable cut apart and/or rolled it to get a rounded cable? G. Unroll it and try B., if cut apart then start with A. Is the drive mounted in a removable drive rack? H. Remove HDD from the drive-bay and start with A. However you should checkout the section in Dansdata's "IDE Fancy Leads, the terrible truth" as to why with all this goin on, for the most part, it still works anyway :p Check out what your chipset has to sort out here http://www.vicstech.com/en/rd3xp/NoiseTest/ click on a picture to see an animated test (note not all types of cables where employed, for instance there are no high quality TPO or teflon cables in this test) like the Power Supply, cables are widely underated as a source of problems, and few ever spend any money on them for anything but "looks" System Timing By setting the memory timings to aggressively, overclocking the Front Side Bus (without locking the PCI bus) or using a hard disk interface transfer mode that is too fast for the system or device (or cables) can cause data loss These days most "enthusiast" boards allow the PCI bus to be locked or employ a divider, but overclocking is always a risk to your data. A more common tweak is aggressive memory timings and those to can effect data intgrity
Quote:
|
|
#4
|
|||
|
|||
|
FileSystem
http://ntfs.com/data-integrity.htm An Explanation of CHKDSK and the New /C and /I Switches <MORE "To understand when it might be appropriate to use these switches (/C and /I) , it is important to have a basic understanding of some of the internal NTFS data structures, the kinds of corruption that can take place, what actions CHKDSK takes when it verifies a volume, and what the potential consequences are in circumventing CHKDSK's usual verification steps. CHKDSK's activity is split into three major "passes" during which it examines all the "metadata" on the volume and an optional fourth pass. Metadata is "data about data." It is the file system overhead, so to speak, that is used to keep track of everything about all of the files on the volume. Metadata tells what allocation units make up the data for a given file, what allocation units are free, what allocation units contain bad sectors, and so on. The "contents" of a file, on the other hand, is termed "user data." NTFS protects its metadata through the use of a transaction log. User data is not so protected. During its first pass, CHKDSK displays a message on the screen saying that it is verifying files and counts from 0 to 100 percent complete. During this phase, CHKDSK examines each file record segment (FRS) in the volume's master file table (MFT). Every file and directory on an NTFS volume is uniquely identified by a specific FRS in the MFT and the percent complete that CHKDSK displays during this phase is the percent of the MFT that has been verified. During this pass, CHKDSK examines each FRS for internal consistency and builds two bitmaps, one representing what FRSs are in use, and the other representing what clusters on the volume are in use. At the end of this phase, CHKDSK knows what space is in use and what space is available both within the MFT and on the volume as a whole. NTFS keeps track of this information in bitmaps of its own that are stored on the disk allowing CHKDSK to compare its results with NTFS's stored bitmaps. If there are discrepancies, they are noted in CHKDSK's output. For example, if an FRS that had been in use is found to be corrupted, the disk clusters formerly associated with that FRS will end up being marked as available in CHKDSK's bitmap, but will be marked as being "in use" according to NTFS's bitmap. During its second pass, CHKDSK displays a message on the screen saying that it is verifying indexes and counts from 0 to 100 percent complete a second time. During this phase, CHKDSK examines each of the indexes on the volume. Indexes are essentially NTFS directories and the percent complete that CHKDSK displays during this phase is the percent of the total number of directories on the volume that have to be checked. During this pass, CHKDSK examines each directory on the volume for internal consistency and also verifies that every file and directory represented by an FRS in the MFT is referenced by at least one directory. It also confirms that every file or subdirectory referenced in each directory actually exists as a valid FRS in the MFT and checks for circular directory references. Finally, it confirms that the various time stamps and file size information associated with files are all up-to-date in the directory listings for those files. At the end of this phase, CHKDSK has ensured that there are no "orphaned" files and that all the directory listings are for legitimate files. An orphaned file is one for which a legitimate FRS exists, but which is not listed in any directory. When an orphaned file is found, it can often be restored to its rightful directory, provided that directory is still around. If the directory that should hold the file no longer exists, CHKDSK will create a directory in the root directory and place the file there. If directory listings are found that reference FRSs that are no longer in use or that are in use but do not correspond to the file listed in the directory, the directory entry is simply removed. During its third pass, CHKDSK displays a message on the screen saying that it is verifying security descriptors and counts from 0 to 100 percent complete a third time. During this phase, CHKDSK examines each of the security descriptors associated with each of the files and directories on the volume. Security descriptors contain information regarding the owner of the file or directory, NTFS permission for the file or directory, and auditing information for the file or directory. The percent complete in this case is the percent of the number of files and directories on the volume. CHKDSK verifies that each security descriptor structure is well formed and internally consistent. It does not verify that the listed users or groups actually exist or that the permissions granted are in any way appropriate. The fourth pass of CHKDSK is only invoked if the /R switch is used. /R is used to locate bad sectors in the volume's free space. When /R is used, CHKDSK attempts to read every sector on the volume to confirm that the sector is usable. Sectors associated with metadata are read during the natural course of running CHKDSK even when /R is not used. Sectors associated with user data are read during earlier phases of CHKDSK provided /R is specified. When an unreadable sector is located, NTFS will add the cluster containing that sector to its list of bad clusters and, if the cluster was in use, allocate a new cluster to do the job of the old. If a fault tolerant disk driver is being used, data is recovered and written to the newly allocated cluster. Otherwise, the new cluster is filled with a pattern of 0xFF bytes. When NTFS encounters unreadable sectors during the course of normal operation, it will also remap them in the same way. Thus, the /R switch is usually not essential, but it can be used as a convenient mechanism for scanning the entire volume if a disk is suspected of having bad sectors. The preceding paragraphs give only the broadest outline of what CHKDSK is actually doing to verify the integrity of an NTFS volume. There are many specific checks made during each pass and several quick checks between passes that have not been mentioned. Instead, this is simply an outline to the more important facets of CHKDSK activity as a basis for the following discussion regarding the time required to run CHKDSK and the impact of the new switches provided in SP4" MORE> Description of Enhanced Chkdsk, Autochk, and Chkntfs Tools in Windows 2000 Pilot Error Well this section is self explanitory, but is actually the leading cause of data loss, through improper data recovery or repair, or simple misadventure like formatting the wrong HDD. Id like to relate how I lost a 200GB RAID 5 array to illustrate the problem in this case was I read the manual :p which stated that the Promise SX6000 wrote array data to the drives and would automatically redetect those drives if they changed channels (of which there are six), so I didnt bother to mark which drives where on which channels when I migrated the array, but when I hooked it back up it only recognized one of the drives, upon further research buried deep in the support website, it had the caution to never removal all the drives from all the channels (from what I understand this issue is no longer a problem but was with that BIOS version) do you know how many possible combinations of six drives to six channels there are? 720 |
|
#5
|
|||
|
|||
|
SATA Cautions
Quote:
http://www.theregister.co.uk/2001/03...en_pc_is_dead/ there is a demo version of ATA Command Test available for download there as well ;) quite a few people have fabricated their own sheilded cables to good effect the cited 10>20% speed decrease, is now out of date with native SATA |
|
#6
|
|||
|
|||
|
Lots of information. Stickied.
__________________
'~"Soli Dio Gloria"~' If you prick us do we not bleed? If you tickle us do we not laugh? If you poison us do we not die? And if your rob us shall we not revenge? あ? なんでもない。。。 http://www.viperlair.com |
|
#7
|
||||
|
||||
|
Wooooaaa, Brain overload... Good information tho (/me checks cable routing of SATA cables).
I take it you got bit by data loss bug :(, hence the great research/Info
__________________
________________ <xterm> The problem with America is stupidity. I'm not saying there should be a capital punishment for stupidity, but why don't we just take the safety labels off of everything and let the problem solve itself? ________________ Brook Moore aka Tw1st3d Kn0t Viperlair Sr. Staff Writer Asus P5W DH Deluxe Intel Xeon 3050 (2.13@3.00) Asus EAX1650XT 2GB SuperTalent PC2-6400 |
|
#8
|
||||
|
||||
|
I am really beginning to like this guy.
__________________
Main Rig: Core 2 Duo E6300, 2 GB Kingston PC6400, MSi 7600GT HD Edition, 80Gb Maxtor, 250Gb Western Digital, 3x 250Gb Western Digitals in RAID 5, 2x 20" widescreen LGs HTPC: A64 3000+, ATi 9700Pro, 2Gb pc3200, 160Gb Seagate, 320Gb Seagate, SI-97 w/ 92mm Panaflo, XBox 360 HD DVD drive, 42" Sharp Aquos 1080P LCD "Let's just pull our d*cks out of our pants and leave the door open, and see who stops by." |
|
#9
|
|||
|
|||
|
Quote:
then I started running various RAID arrays (most levels) ended up with enough drives to brownout the local grid, and then I got tapped as the Moderator for Data Storage at [H]ard|Forums ![]() so I just had this laying around and since I mentioned Corruption in a different thread... thank you for the sticky ![]() Last edited by Ice Czar : 05-18-2004 at 06:49 PM. |
|
#10
|
|||
|
|||
|
I'll toss a bit more in
As the Hard Disc Spins indepth analysis of HDD performance @ Lost Circuits Summary Excerpts "Hard disk drive technology has moved more and more into the center of attention of the IT industry. The original role of HDDs was meant to be a simple mass storage media with relatively little emphasis on performance; twenty years ago, this did suffice. With the internet evolving and servers becoming a powerful factor in the electronic fabric of data communication throughout the entire world, it soon became obvious that storage media were also the most prevalent bottleneck, thus, the need for increased performance. Performance of hard disk drives is a rather touchy topic, depending on who one asks, one may get very different answers, based on the different criteria applied and the different benchmarks used. Buzz words like internal performance, seek latencies, media transfer rates, response times are being thrown around, burst speed and effective host transfer rates are held against those and last not least, the cache size matters. This is the level of argumentation for a single drive, once the subject switches to RAID configurations, things are getting even more complicated. In addition, there are the standard benchmarks and then there are those advanced benchmarks like IOMeter, which by their mere use already qualify the testers as experts. Or not? In this and the following articles we will first discuss HDD performance in general, then go about a few basic functional drive architecture parameters and finally discuss how certain benchmarks fit into the grand scheme. We will also show, which benchmarks, despite their popularity are either meaningless for the end user or else create a false impression of the drive's capabilities. There will be follow-up articles that will deal in greater detail with individual benchmarks. " I: Internal Drive Performance What is HDD Performance Anyway? Hard Disk Drive Architecture Zones Effective Internal Transfer Rate (TxD) Internal Performnance vs Sequential Data Transfer Rate Servo Bursts Skew Summary: Effective Internal Performance (TxD) "Effective internal performance or effective media transfer rate is defined by the media density times the linear velocity relative to the head minus housekeeping data that are interspersed with the actual data on the platters. Higher area densities and higher rotational speeds will increase the effective internal performance, higher amounts of housekeeping data as required for e.g. positional corrections will reduce the effective transfer rate. Depending on the drive's targeted environment, the house-keeping overhead will vary, e.g. laptop drives that are subject to higher levels of vibrations will need more "corrective actions" / repositioning than desktop drives. Likewise, rack-mounted server drives may require more servo data since vibrations can easily propagate throughout an entire rack. It should also be clear now why the relatively higher amount of servo data on high-end, e.g. SCSI drives will result in lower performance than that of comparable (in terms of rpm) desktop drives. On the other hand, within the environment that SCSI drives usually are operating in, the very same "faster" desktop drives would run into the need for constant recalibration which would cause a severe performance hit under operational conditions" II: Averages, Seeks and other Paradoxes Average Sequential Transfer Read vs. Write Performance Random Access Long Seek vs. Short Seek Optimizations Longer Platters Will Cause Higher Seek/ Random Access Latencies Summary "In this (second) article, we have covered some misleading internal drive parameters like the average sequential transfer rates for READ and WRITE as well as the different parameters influencing Random Access and Seek latencies and we further showed that different capacity models of the same drive can generate some performance data that paint the opposite picture of what the user would get in real world performance. Keep in mind that all parameters covered so far are only indicative of the drive's internal performance and should not vary from one interface to the other, at least in single drive test configurations. That means that e.g. using the "average sequential transfer" to show differences in interfaces or controllers is somewhat useless since by definition the results will be the same. Comparing e.g. the SATA and PATA interface on the same setup with different drives can only be called a self-fulfilling prophecy and will not bear any relevance for the intended test either. In the case of RAID setups, there will be differences between READ and WRITE performance, moreover, the interface bandwidth will likely become a limiting factor and the cause for speed matching conditions. " III: Effective Host Transfer Rates Interface Speed as Basis for Classification A Brief History of ATA Host Transfer Rates (TH) and Effective Host Transfer Rates (TxH) Quantifying the CMD Overhead Data Transfer Vs. I/O Performance Summary "In the last two articles, we have covered the internal performance parameters of hard disc drives, that is, more precisely, those parameters that mostly relate to a HDD as an electromechanical device. For a brief recap, HDDs use electronic and mechanical parts and the mechanical latencies are what mostly holds back the performance of any disc drive. The main reason is the inertia of any mechanical component, starting from the spindle and the platter to the read / write heads and the actuator. Whoever read through the first two articles on the subject should at least have some basic understanding of how media density and rotational speed translate into sequential transfers, likewise, the influence of rotational latency and seek latency on the random access speed should be clear" IV: DMAs, Latencies and Speed Matching Bus Master? Other factors influencing host transfer rate ..PCI Latency ..Bus Parking ..Write Combining Speed Matching Condition Summary "In the last article, I briefly brushed on the issue that current version of Windows, as well as Unix and Linux systems are not real time operating systems, rather, software stacks are created to allow internal scheduling of the execution of events. That, however, also means that there can be conflicts between the availability of interrupts and the software stacks, that is, the hardware is fighting against the software for priority. The result in cases like that is that there is no winner, performance is being reduced and at the same time, CPU usage is increased because of access errors and retries on the PCI bus and ineffective execution of transfers. Some of the issues with poor utilization are plain and simply a matter of incorrect or marginally functional software. One good example are the chipset and bus master drivers that are necessary for almost any chipset not conforming to the original Intel specifications. That is, there are hardware bridges and buffers that need to be configured in order to optimally interface with the Microsoft OS environment by means of drivers. Only in a few rare instances will the new interface be completely transparent to the OS, one example is the ICH5 Serial ATA interface introduced by Intel with the Canterwood / Springdale chipset. However, even in this case, the transparency (the fact that the OS does not even notice anything has changed) is limited to the non RAID version of the same south bridge, namely the ICH5R, in that for RAID operation the installation of drivers will be necessary." "The VIA Hyperion drivers are the currently last step in a development that we have followed for about 5 years to optimize the South Bridge and IDE controller interaction with the Windows environment by means of bus master and GART drivers." "Third Party chipsets" and external controllers will in almost all cases require the installation of extra drivers, not necessarily for any basic functionality but definitely for enabling performance while reducing CPU utilization. A few well known examples are the VIA or nVidia or any other chipset bus master drivers." V: Protocol Differences For Reduced Latencies Cabling and Parallel Signaling Properties Serial ATA Shared Bus Issues Parallel vs. Serial Command Overhead Serial ATA: First Party DMA SATA "Out Of Order Data Delivery" To Reduce Rotational Latencies "we have concentrated on some of the internal design parameters that influence the internal performance of Hard Disc Drives as well as some of the issues that relate to the effective host transfer rate and the interfacing of the drive with the Host Bus Adapter (HBA) and the DMA and busmastering channels needed to interface with the system logic. The current article will concentrate more on the differences between Parallel and Serial ATA with a focus on the different interfacing protocols Most of the marketing strategies promoting the migration from Parallel to Serial ATA have focused on the physical cabling properties, that is, the reduction from 16 bidirectional data channels to two unidirectional pairs of Low Voltage Differential Signaling ((LVDS) lines. On the surface, the obvious effect is a greatly facilitated ease of routing, reduced obstacles in the air flow and a smaller connector footprint. The main reasons, however, relate to the signaling properties in the context of the fact that parallel signaling across long distances had no headroom left for further speed grades. " |
|
#11
|
|||
|
|||
|
VI: Command Queing
Queuing Schemes: Parallel ATA vs.Serial ATA Mechanical Overhead Seek Latencies Supermarkets and Elevators Rotational Latencies Different Queues in Different Standards The Hardware Behind Queuing Queue Depth against The Rest Of The World The Big Picture In summary, here is the short and sweet on the different forms of Command Queuing. In Parallel ATA, the merely passive role of the drive, along with the command overhead associated with the disconnect and polling of master and slave devices on the same cable renders the legacy command queuing scheme somewhat ineffective. Therefore, there has been little incentive to move to the more sophisticated albeit more expensive command queuing scheme. Serial ATA extensions add Native Command Queuing to the FirstPartyDMA engine setup. The point to point topology allows a continuous communication between the device and the controller, which in turn allows to take full advantage of advanced features like for example a so-called non-zero offset DMA engine setup to allow for out of order data delivery, as well as reordering of commands within the queue. SATA NCQ does not allow prioritizing of queues, however, Virtual Head of Queue attributes are possible with the effect that any such command will trash any existing queue. This virtual Head of Queue command will, thus, grant priority to itself in a Last-Man-Standing fashion. Thereafter, the previously uncompleted commands will have to be reissued. RAID I: The Lesser Levels Summary MOST OF OUR PREVIOUS ARTICLES have focused on single drive technology, that is, HDD architecture, interface protocols and the back-end support of ATA by means of DMA channels. Likewise, we have shown some issues with the use of a number of benchmarks that can yield false results caused by speedmatching conditions. Especially the latter issue mostly occurs with RAID configurations, reason enough to start a short RAID overview series. For starters, we will cover the "lesser" forms of RAID, that is RAID Level0, 1 and 10, before taking the plunge into Exclusive-Or calculations that form the operational backbone of real RAID configurations Different ?Categories? of RAID In general, on we need to distinguish between three different categories of RAID, namely: 1. Standalone RAID solutions for mass storage in the backplane of servers (fiber channel-attached) or Network-Attached Storage (NAS) connected via Firewire or Gigabit Ethernet. 2. RAID functionality via separate Host Bus Adapter (HBA) cards using PCI, PCI-X (64bit / 66 MHz) or PCI-Express (3GIO) interface. 3. RAID controller integrated on the mainboard level. Often called RAID-lite because of limited functionality, except for dedicated Server boards. RAID on Mainboard (ROMB). Software vs. Hardware RAID Software RAID Hardware RAID JBOD and Spanning RAID Level1 Mirroring / Duplexing Duplexing vs. Mirroring RAID Level 0+1 (Level 10) As a summary of where we are and what is about to come in the next few articles, currently used levels of RAID are: Lesser RAID Levels: Level 1: Mirroring Level 1: Duplexing Level 0: Striping Some companies use the term "spanning" when they really mean striping. Spanning normally only refers to JBOD Level 1/0 or 10: Mirroring of Striped Drives (Expensive!!!) and next True RAID Levels Level 2: Bit-Level striping across a minimum of 11 drives using Hamming codes, a form of error correcting code (ECC) Level 3: Byte Level striping with Parity Level 4: Block Level Striping with dedicated Parity Level 5: Block Level Striping with distributed Parity Level 6: Same as Level 5 but with dual distributed Parity Level 7: Multiple Cache-level striping with dedicated Parity (proprietary format of Storage Computer Corporation including real-time processor) We will go through those step by step in the next few articles. |
|
#12
|
|||
|
|||
|
Advanced HDD Issues Linkfram
Hard Disk Drive Reference Section @ Storagereview.com (reprinted from the PC Guide link below) Including: A Brief History of the Hard Disk Drive Construction and Operation of the Hard Disk Drive Hard Disk Geometry and Low-Level Data Structures Hard Disk Performance, Quality and Reliability Hard Disk BIOS and Capacity Factors Hard Disk Interfaces and Configuration Hard Disk Logical Structures Hard Disk Drives A the PC Guide (same as above) PC Guide Topic Index ATA-ATAPI.COM How it Works Document series (HTML and Zip available) Including: Hale's ATA FAQ Fact and Fiction CHS Translation Partition Tables Masterboot Record DOS Floppy Disk Boot Sector OS2 Boot Sector Computer Boot Sequence @ Mossywell.com Including: Hard Disk Geometry How the Physical Disk is accessed The Standards How we used to access the disk: CHS, ECHS, Revised ECHS, Assisted LBA, LBA How we now access the disk: LBA and Extended INT13h Better than LBA and Extended INT13h: Direct Disk Access (DMA) The BIOS The Master Boot Record \ The Master Boot Record Code Partition Boot Sector and Clusters \ Partition Boot Sector Code FAT Locations \ The FAT in Detail The Root Directory \ The Root and Other Directories in Detail IO.SYS with MS-DOS NTLDR with Win NT (W2K, XP) NTFS vs FAT The NTFS Filesystem Comprehensive Overview Fat System Guide NTFS Basics Converting FAT32 to NTFS (you should really read this) Default Cluster Sizes (chart) Windows 2000 and the Boot @ Windows & .NET Magazine Inside the Boot Process Part 1 (NTFS) & Part 2 @ Windows & .NET Magazine Partition Strategies @ Radified.com Fdisk Guide @ Radified.com Bootdisks Ultimate Boot CD A MUST HAVE Bootdisk.com ETPlanet TCP/IP bootdisks + Ultimatebootcd Bootable cd image with: Hard Disk Diagnosis, Drive Fitness Test (IBM/Hitachi) 3.50, PowerMax (Maxtor/Quantum) 4.06, Data Lifeguard (Western Digital) 10.0, SeaTools Desktop (Seagate) 1.06.02, Diagnostic Tool (Fujitsu) 6.10, SHDIAG (Samsung) 1.25, Hard Disk Management, IBM/Hitachi Feature Tool 1.90, Ranish Partition Manager 2.43, AutoClave (HDD Wiper) 0.3, Partition Resizer 1.3.4, SavePart (Partition Saver) 2.70, XFDISK (Extended FDISK) 0.9.3beta, g4u (HDD Cloning) 1.12, HDClone (Free Edition) 1.0, TestDisk 4.4, Memory Diagnosis, Memtest86 3.0, Linux-based Rescue Disks, Offline NT Password & Registry Editor 030426, Tom's Boot Disk 2.0.103, Recovery Is Possible (RIP) 2.0, RIP Linux Rescue Disk, and AIDA16 (System information) 2.08, F-Prot Antivirus for DOS (Personal use only), Virus definition: 22 Aug 2003, Macro virus definitions have been disabled so that everything can fit onto a 2.88MB virtual floppy boot image. Includes read-only freeware version of NTFSDOS 3.14a, Active NTFS Reader for DOS, Thanx to styckx who originally posted this Boot Managers Gujin Understanding MultiBooting and Booting Windows from an Extended Partition XOSL opensource freeware Smart Boot Manager opensource freeware System Commander $ OS-BS FreeBSD boot manager opensource freeware Ranish Partition Manager freeware GNU GRUB opensource freeware LILO Linux Bootmanager opensource freeware Solaris boot manager Masterbooter shareware Linux-NTFS Project Microsoft Disk Reference How Windows 2000 Assigns, Reserves, and Stores Drive Letters Diskpart Utility HOW TO: Change Drive Letter Assignments in Windows XP Additional Reference Serial ATA (SATA) ATA EIDE ATAPI-ATA- EIDE History EIDE vs SCSI SCSI FAQ SCSI FAQ Fibre Channel List of Partition IDs Error codes for Ghost Performances issues and Tradeoffs in Configuring for multiple devices Independent Master Slave Timing Dynamic Disks Description of Disk Groups in Windows Disk Management Dynamic vs. Basic Storage in Windows 2000 Basic and Dynamic Disks @ Windows & .net Magazine HOW TO: Recover an Accidentally Deleted NTFS or FAT32 Dynamic Volume Dynamic Disk Hardware Limitations (No firewire, USB, removable or laptop) HOW TO: Set Up Fault-Tolerant Sets on Dynamic Disks in Windows 2000 Dynamic Disk Numbering and the DmDiag.exe Tool HOW TO: Regenerate a Dynamic Mirrored Volume in Windows 2000 Restrictions on Extending or Spanning Simple Volumes on Dynamic Disks Limits of Dynamic Disks in Windows 2000 LDMDump (Freeware utility) @ sysinternals LDM Database @ Linux-NTFS project LDM FAQ @ Linux-NTFS project Recovery Reference Recovering NTFS Boot Sector on NTFS Partitions |
|
#13
|
|||
|
|||
|
NTFS Advanced Studies
NTFS Volume Management and HKLM\System\DiskKey NTFS Boot INI Options Reference NTFS Defragmenting Inside W2K NTFS Part 1 Inside W2K NTFS Part 2 Exploring NTFS On-Disk Structures Inside Storage Management, Part 1 Inside storage Management Part 2 Basic vs Dynamic Disks Inside Encrypting File System Part 1 Inside Encrypting File System Part 2 Inside Memory Management Paging Files Additional NT Articles at Windows & .NET Magazine By Mark Russinovich Including: Crash Dump Analysis, Inside Win32 Services, Windows 2000 Kernal, Scalability Enhancements, Management Interface, Reliability Enhancements, and the Registry. Additional NTFS Articals and Utilities atSysinternals Other Filesystems The Linux Filesystem Explained Linux Filesystems Comparison Ext2FS Ext3 ReiserFS JFS for Linux The Unix UFS Filesystem Space efficiency SFS,FFS,AFS,FAT16,FAT32 RAID Definitive Guide to RAID @ Storagereview RAID I: The Lesser Levels @ Storage Review (0, 1 mirroring, 1 duplexing, 1+0) RAID an In-Depth Guide @ SLCentral.com The Skinny on RAID @ arstechnica RAID: Your Guide @ PCMechanic RAID Explained @ AnandTech (part of IDE RAID Comparison dated) SATA Serial ATA in the Microsoft Operating System Environment including: The Significance of Serial ATA The Different Modes of Serial ATA Controllers Serial ATA 1.0 Features and Details Emulating Parallel ATA Mode Native Serial ATA Mode SATA II Features and Details Serial ATA Hardware Register Interface Naming Conventions for Serial ATA Products Support for Serial ATA in Windows Ataport Serial ATA Emulating Parallel ATA Mode Controller Support in the Windows Family of Operating Systems Native Serial ATA Mode Controller Support in Future Versions of Windows Emulating Parallel ATA Mode Controller Program Identifying Emulating Parallel ATA Mode and Native Serial ATA Mode Controllers Multiple Controllers in a System Booting from the Different Modes of Serial ATA Serial ATA as an External Connection Hot Plugging Hard Disk Drive Capacity Limitations on Serial ATA CD-ROM Opportunities and BIOS Settings for Native-Mode-Capable ATA Controllers Defragmentation O&O Defrag Pro PageDefrag Freeware Utilities BootPart+ PartitionInfo and Partition Table and Boot Record Editor Savepart Partition Resizer Zpart Partition Image for Linux Partition Image is a Linux/UNIX utility which saves partitions in many formats (see below) to an image file. The image file can be compressed in the GZIP/BZIP2 formats to save disk space, and split into multiple files to be copied on removable floppies (ZIP for example), ... Partitions can be saved across the network since version 0.6.0. GNU Parted a program for creating, destroying, resizing, checking and copying partitions, and the file systems on them. Can be run from a GNU\Linux boot image supports: ext2, ext3, fat16, fat32, linux-swap, HFS, JFS, NTFS, ReiserFS, UFS, XFS FIPS is a program for non-destructive splitting of harddisk partitions. TestDisk a tool to check and undelete partitions Disk Utilities THE LIST more freeware disk utlities than you can shake a stick at Diskmon Freeware a Windows NT device driver/GUI combination that together log and display all process activity on a Windows NT/2000 system. You can also minimize Diskmon to your system tray where it acts as a disk light, presenting a green icon when there is disk-read activity and a red icon when there is disk-write activity. Filemon Freeware Monitors and displays file system activity on a system in real-time NTFSInfo Freeware Information about NTFS volumes. Dump includes the size of a drive's allocation units, where key NTFS files are located, and the sizes of the NTFS metadata files on the volume. Disk Investigator Freeware View and search raw directories, files, clusters and system sectors. ActiveSmart Trialware S.M.A.R.T. diagnostic and failure prediction software for hard drives AIDA32 Freeware S.M.A.R.T. Monitoring, Drive, ASPI, ATA Info, Plus lots lots more DBAN (Dariks Boot and Nuke a self-contained boot floppy that securely wipes the hard disks of most computers submitted by roncomatic File \ Directory Utilities i.disk freeware SequoiaView freeware Flash Utilities MtkWinFlash {freeware) Windows utility will flash most Mediatek-chip based ATAPI drives supports all single-file firmware in BIN and HEX format (see link for supported models includes most LiteOn and many more) -Contributed by TechHead Bask Backup GHOST Guide to backing up with Ghost @ Radified.com Drive Image XXCopy Storebackup (for Linux, GNU freeware) HDClone Free Edition (rudimentary Clone Utility) g4u both local and via ftp clone utility (Freeware BSD license) all filesystems supported ASPI ASPI Layers/Drivers @ Radified.com ACPI HAL http://support.microsoft.com/default...;en-us;Q246236 |
|
#14
|
|||
|
|||
|
Benchmarking
IOMeter Freeware (Open Source) IOMeter User Guide Introducing IOMeter @ StorageReview (Testbed II) Using IOMeter @ 2CPU.com (contributed here by big daddy fatsacks) Note from the Storage Review FAQ "Can IOMeter measure single-user performance? In a nutshell, no it can not. With all due respect to both Intel and to the SourceForge.net team (cache) that picked up the project after Intel discontinued it, IOMeter measures random access performance across varying loads and simulated nodes (network performance). In otherwords, while it remains quite suited to assessing multi-user performance, it has little capability to tackle single-user scenarios. Why not? As explained on a page in our comprehensive methodology outline (cache), IOMeter does not have the capability to accurately simulate the localized data access that dominates single-user drive use. That is, it can not simulate the tendancy for a drive to spend a large amount of time seeking across a very small area relative to broader-stroke movements. StorageReview? itself is primarily to blame for IOMeter's popularity across hardware sites to simulate "Workstation" usage. One of the goals of the Testbed3 project was to recant this erroneous deployment while introducing far more accurate tools to assess single-user (desktop/workstation) performance. It has been nearly two years... we hope that other hardware evaluation sites will eventually give careful consideration to the theories behind IOMeter's inability to simulate locality and as a result consider removing the "Workstation" pattern from their performance suites." IOZone Freeware Important Filesystem Benchmark WinBench 99 (freeware) ATTO HD Tach Diskspeed32 Freeware George's HDSpeed Disk Performance Test Freeware Rudimentary Interface speed and Read Sequential Sustained Transfer Rate Disk Bench Bart's SCSItool IPEAK SPT - Intel Performance Evaluation & Analysis Kit Storage Performance Toolbox (StorageReview's new Testbed 3) yours for a measly $995 :eek: which is actually a suite of utilities WinTrace32, AnalyzeTrace, AnalyzeLocality, RankDisk and AnalyzeDisk IPEAK SPT AnalyzeDisk Reference Guide @Storage Review,com a good read if nothing else ;) The Older StorageReview Testbeds Testbed I Testbed II Optical Testbed Legacy Benchmark Database RAID.edu Benchmarks many Unix NT and DOS Benchmarks (some unavailable anywhere else) UNIX Bonnie++ Bonnie v.2.0.6 IOZone IOStone v. C/II Disktest IOBench IOCall RawIO PostMark IOGEN DOS SCSITool RAIDmark Qbench COREtest NT\W2K|XP Nbench NTiogen Threadmark (others are already linked above) I'll see what I can do to whack this into better shape as I see several formating changes, omissions, spelling errors and a few bad links ;) Last edited by Ice Czar : 05-18-2004 at 07:11 PM. |
|
#15
|
||||
|
||||
|
daaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaammmm
i'll do a search command on this thread everytime i need comp info now.
__________________
Come on baby turn the lights off let's get naked Come on babygirl you know I just can't take it Let your panties hit the floor Let your body steal the show right now |
|
|
|
| Thread Tools | |
| Display Modes | |
|
|
|
|