View previous topic :: View next topic |
Author |
Message |
R0b0t1 Apprentice
Joined: 05 Jun 2008 Posts: 264
|
Posted: Fri Dec 24, 2021 6:28 am Post subject: |
|
|
I don't mean to hijack the conversation, but w.r.t. the topic I'm in need of some help figuring out a reliable way to get drive life. I was originally going to comment that you can't actually know because most drives hide failure rates[1] but realized I should check my drives again. Having checked, this seems true. Only one of them directly reports drive lifespan left but it seems to be inaccurate and always 100. Generally I would make an observation like OP did and conclude the drive is failing but not telling me about it.
This also opens up some questions about how LVM organizes writes. The first disk in my primary volume group has received 7tb of traffic, ~3.5x the next disk in the volume group.
Code: |
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x000b 100 100 050 Pre-fail Always - 2
9 Power_On_Hours 0x0012 100 100 000 Old_age Always - 36335
12 Power_Cycle_Count 0x0012 100 100 000 Old_age Always - 72
168 SATA_Phy_Error_Count 0x0012 100 100 000 Old_age Always - 0
170 Bad_Blk_Ct_Erl/Lat 0x0003 075 075 010 Pre-fail Always - 52/216
173 MaxAvgErase_Ct 0x0012 100 100 000 Old_age Always - 30 (Average 5)
192 Unsafe_Shutdown_Count 0x0012 100 100 000 Old_age Always - 69
194 Temperature_Celsius 0x0023 070 070 030 Pre-fail Always - 30 (Min/Max 30/30)
218 CRC_Error_Count 0x000b 100 100 050 Pre-fail Always - 0
231 SSD_Life_Left 0x0013 100 100 000 Pre-fail Always - 100
241 Lifetime_Writes_GiB 0x0012 100 100 000 Old_age Always - 7468
SMART Attributes Data Structure revision number: 16
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE
1 Raw_Read_Error_Rate 0x000a 100 100 000 Old_age Always - 0
2 Throughput_Performance 0x0005 100 100 050 Pre-fail Offline - 0
3 Unknown_JMF_Attribute 0x0007 100 100 050 Pre-fail Always - 0
5 Reallocated_Sector_Ct 0x0013 100 100 050 Pre-fail Always - 0
7 Unknown_JMF_Attribute 0x000b 100 100 050 Pre-fail Always - 0
8 Unknown_JMF_Attribute 0x0005 100 100 050 Pre-fail Offline - 0
9 Power_On_Hours 0x0012 100 100 000 Old_age Always - 37091
10 Unknown_JMF_Attribute 0x0013 100 100 050 Pre-fail Always - 0
12 Power_Cycle_Count 0x0012 100 100 000 Old_age Always - 72
167 Unknown_JMF_Attribute 0x0022 100 100 000 Old_age Always - 0
168 SATA_Phy_Error_Count 0x0012 100 100 000 Old_age Always - 70
169 Unknown_JMF_Attribute 0x0013 100 100 010 Pre-fail Always - 262245
170 Bad_Block_Count 0x0013 100 100 010 Pre-fail Always - 0 0 0
173 Erase_Count 0x0012 180 180 000 Old_age Always - 22 476 344
175 Bad_Cluster_Table_Count 0x0013 100 100 010 Pre-fail Always - 0
192 Unexpect_Power_Loss_Ct 0x0012 100 100 000 Old_age Always - 58
194 Temperature_Celsius 0x0022 073 073 030 Old_age Always - 27 (Min/Max 25/40)
197 Current_Pending_Sector 0x0012 100 100 000 Old_age Always - 0
240 Unknown_JMF_Attribute 0x0013 100 100 050 Pre-fail Always - 0
SMART Attributes Data Structure revision number: 1
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE
5 Reallocated_Sector_Ct 0x0033 100 100 010 Pre-fail Always - 0
9 Power_On_Hours 0x0032 092 092 000 Old_age Always - 37209
12 Power_Cycle_Count 0x0032 099 099 000 Old_age Always - 67
177 Wear_Leveling_Count 0x0013 099 099 000 Pre-fail Always - 10
179 Used_Rsvd_Blk_Cnt_Tot 0x0013 100 100 010 Pre-fail Always - 0
181 Program_Fail_Cnt_Total 0x0032 100 100 010 Old_age Always - 0
182 Erase_Fail_Count_Total 0x0032 100 100 010 Old_age Always - 0
183 Runtime_Bad_Block 0x0013 100 100 010 Pre-fail Always - 0
187 Uncorrectable_Error_Cnt 0x0032 100 100 000 Old_age Always - 0
190 Airflow_Temperature_Cel 0x0032 075 058 000 Old_age Always - 25
195 ECC_Error_Rate 0x001a 200 200 000 Old_age Always - 0
199 CRC_Error_Count 0x003e 100 100 000 Old_age Always - 0
235 POR_Recovery_Count 0x0012 099 099 000 Old_age Always - 65
241 Total_LBAs_Written 0x0032 099 099 000 Old_age Always - 4149305949
|
The middle one is an ADATA. We've got interesting things like SATA_Phy_Error_Count but nothing I can link to total writes. At least there's some actual failure stats in there?
Last one is an 850 EVO. Sector size is 512, ~2TB written.
[1]: If you have two drives which are nominally the same, but one of which reports accurate failure rates, it is perceived to be worse because it's not lying. So all manufacturers seem to make lying drives. |
|
Back to top |
|
|
NeddySeagoon Administrator
Joined: 05 Jul 2003 Posts: 54810 Location: 56N 3W
|
Posted: Fri Dec 24, 2021 7:56 am Post subject: |
|
|
Split from SSD deterioration, as its a different topic
R0b0t1,
You may be looking in the wrong place. Please put all of onto a pastebin.
The first one says
Code: | 241 Lifetime_Writes_GiB 0x0012 100 100 000 Old_age Always - 7468 |
The last one says
Code: | 241 Total_LBAs_Written 0x0032 099 099 000 Old_age Always - 4149305949 |
Divide those by the size of the drive to get average drive erase cycles.
Then you have used about that much of the rated erase cycles of the drive life.
QLC drives are about 800 erase cycles.
The Code: | 231 SSD_Life_Left 0x0013 100 100 000 Pre-fail Always - 100 | is not useful.
Thee 100 raw value is in unknown units.
The normalised is not a lot of use either. How do you know its linear? _________________ Regards,
NeddySeagoon
Computer users fall into two groups:-
those that do backups
those that have never had a hard drive fail. |
|
Back to top |
|
|
eccerr0r Watchman
Joined: 01 Jul 2004 Posts: 9884 Location: almost Mile High in the USA
|
Posted: Fri Dec 24, 2021 8:32 pm Post subject: |
|
|
Speaking of your LVM, how are you using it in terms of how you created the logical volumes? Whether special features like LVMRAID was used or whether you simply did linear append, it will determine how the logical volumes are allocated. I believe the default "first" physical volume that's used is what you specified first when you created the volume group, but you can override this when you actually create the logical volumes.
Because LVM extents can be arbitrarily allocated much like a regular filesystem, it can also be subject to fragmentation...
Then after your LV is created, it follows the same rules of how the underlying filesystem (ext4fs or whatever you're using) allocates blocks which is a completely different issue. _________________ Intel Core i7 2700K/Radeon R7 250/24GB DDR3/256GB SSD
What am I supposed watching? |
|
Back to top |
|
|
|
|
You cannot post new topics in this forum You cannot reply to topics in this forum You cannot edit your posts in this forum You cannot delete your posts in this forum You cannot vote in polls in this forum
|
|