[GH-ISSUE #11] Not working for rs3621xs+ on DSM 6.2.4-25556 #12

Closed
opened 2026-03-07 19:13:56 +03:00 by kerem · 33 comments
Owner

Originally created by @nightmadeoflight on GitHub (Mar 17, 2023).
Original GitHub issue: https://github.com/007revad/Synology_HDD_db/issues/11

Drives added (and now showing as "already existing"):
WD4002FFWX-68TZ4N0 already exists in rs3621xs+_host.db
WD4002FFWX-68TZ4N0 already exists in rs3621xs+_host.db.new
WD40EFRX-68WT0N0 already exists in rs3621xs+_host.db
WD40EFRX-68WT0N0 already exists in rs3621xs+_host.db.new

Before the reboot., I get notifications on the Synology web interface to say they are unverified (each time you run the script).
After a reboot, they are still showing unverified.

Originally created by @nightmadeoflight on GitHub (Mar 17, 2023). Original GitHub issue: https://github.com/007revad/Synology_HDD_db/issues/11 Drives added (and now showing as "already existing"): WD4002FFWX-68TZ4N0 already exists in rs3621xs+_host.db WD4002FFWX-68TZ4N0 already exists in rs3621xs+_host.db.new WD40EFRX-68WT0N0 already exists in rs3621xs+_host.db WD40EFRX-68WT0N0 already exists in rs3621xs+_host.db.new Before the reboot., I get notifications on the Synology web interface to say they are unverified (each time you run the script). After a reboot, they are still showing unverified.
kerem closed this issue 2026-03-07 19:13:57 +03:00
Author
Owner

@007revad commented on GitHub (Mar 18, 2023):

Can you please attach get a copy of your rs3621xs+_host.db file:

sudo cp /var.defaults/lib/disk-compatibility/rs3621xs+_host.db $HOME/rs3621xs+_host.db.txt

Then reply here and attach the rs3621xs+_host.db.txt file.

If you don't have the home folder service enabled change $HOME in the command above to a path you have access to.

<!-- gh-comment-id:1474662696 --> @007revad commented on GitHub (Mar 18, 2023): Can you please attach get a copy of your rs3621xs+_host.db file: `sudo cp /var.defaults/lib/disk-compatibility/rs3621xs+_host.db $HOME/rs3621xs+_host.db.txt` Then reply here and attach the rs3621xs+_host.db.txt file. If you don't have the home folder service enabled change $HOME in the command above to a path you have access to.
Author
Owner

@007revad commented on GitHub (Mar 22, 2023):

OP not responding

<!-- gh-comment-id:1480347233 --> @007revad commented on GitHub (Mar 22, 2023): OP not responding
Author
Owner

@nightmadeoflight commented on GitHub (Mar 23, 2023):

I upgraded DSM to the latest version 7 (from 6), ran the script and it
worked. :)

On Wed, Mar 22, 2023 at 10:38 PM 007revad @.***> wrote:

Closed #11 https://github.com/007revad/Synology_HDD_db/issues/11 as
completed.


Reply to this email directly, view it on GitHub
https://github.com/007revad/Synology_HDD_db/issues/11#event-8821783984,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/A6RXYJKEOYEOMRGCJ5ZESI3W5N5PPANCNFSM6AAAAAAV6RAHNU
.
You are receiving this because you authored the thread.Message ID:
@.***>

<!-- gh-comment-id:1481331576 --> @nightmadeoflight commented on GitHub (Mar 23, 2023): I upgraded DSM to the latest version 7 (from 6), ran the script and it worked. :) On Wed, Mar 22, 2023 at 10:38 PM 007revad ***@***.***> wrote: > Closed #11 <https://github.com/007revad/Synology_HDD_db/issues/11> as > completed. > > — > Reply to this email directly, view it on GitHub > <https://github.com/007revad/Synology_HDD_db/issues/11#event-8821783984>, > or unsubscribe > <https://github.com/notifications/unsubscribe-auth/A6RXYJKEOYEOMRGCJ5ZESI3W5N5PPANCNFSM6AAAAAAV6RAHNU> > . > You are receiving this because you authored the thread.Message ID: > ***@***.***> >
Author
Owner

@Greg-ATG commented on GitHub (Apr 13, 2023):

The issue here may be that some NAS models running DSM6 still reference the "X_v7.db" files. Actually, now that I think about it, our NAS (RS2421RP+) shipped with DSM7, but we completed a downgrade procedure to DSM6. Perhaps this is why it's referencing the v7 databases.

This would explain why @nightmadeoflight upgrading their NAS to DSM7 fixed the issue, since then this script would have started fixing the v7 database files.

Perhaps it's safest to modify the v7 database files (if present) regardless of the DSM version?

<!-- gh-comment-id:1507275691 --> @Greg-ATG commented on GitHub (Apr 13, 2023): The issue here may be that some NAS models running DSM6 still reference the "X_v7.db" files. Actually, now that I think about it, our NAS (RS2421RP+) shipped with DSM7, but we completed a downgrade procedure to DSM6. Perhaps this is why it's referencing the v7 databases. This would explain why @nightmadeoflight upgrading their NAS to DSM7 fixed the issue, since then this script would have started fixing the v7 database files. Perhaps it's safest to modify the v7 database files (if present) regardless of the DSM version?
Author
Owner

@007revad commented on GitHub (Apr 17, 2023):

@Greg-ATG

I've only seen X.db files in DSM 6 and X_v7.db files in DSM 7... but there could be situations where someone's Synology has both. I know when you upgrade from DSM 6 to DSM 7 it appends _v7 to all the .db file names.

Since you have downgraded from DSM 7 to DSM 6 I'm curious if your RS2421RP+ now has both X.db and X_v7.db files? Can you run the following command to see what it returns:

ls /var/lib/disk-compatibility/*.db

<!-- gh-comment-id:1510907772 --> @007revad commented on GitHub (Apr 17, 2023): @Greg-ATG I've only seen X.db files in DSM 6 and X_v7.db files in DSM 7... but there could be situations where someone's Synology has both. I know when you upgrade from DSM 6 to DSM 7 it appends _v7 to all the .db file names. Since you have downgraded from DSM 7 to DSM 6 I'm curious if your RS2421RP+ now has both X.db and X_v7.db files? Can you run the following command to see what it returns: `ls /var/lib/disk-compatibility/*.db`
Author
Owner

@Greg-ATG commented on GitHub (Apr 17, 2023):

Yes, the disk-compatibility directory contains both X.db and X_v7.db files for us. This must be related to the DSM6 downgrade.

<!-- gh-comment-id:1512030286 --> @Greg-ATG commented on GitHub (Apr 17, 2023): Yes, the disk-compatibility directory contains both X.db and X_v7.db files for us. This must be related to the DSM6 downgrade.
Author
Owner

@Greg-ATG commented on GitHub (Apr 19, 2023):

With the "removable disk" issue fixed in the latest version (2.1.38), I've confirmed that running the script on our RS2421RP+ w/ DSM6 does NOT fix the "unverified drives" warning in DSM. However, by hardcoding "_v7" into the script as shown below, running it DOES fix the warnings.
image

Would it be harmful to have the script look for ANY "${model}_host.db(.new)" OR "${model}_host_v7.db(.new)" files and patch all of them, regardless of DSM version installed?

<!-- gh-comment-id:1515248478 --> @Greg-ATG commented on GitHub (Apr 19, 2023): With the "removable disk" issue fixed in the latest version (2.1.38), I've confirmed that running the script on our RS2421RP+ w/ DSM6 does NOT fix the "unverified drives" warning in DSM. However, by hardcoding "_v7" into the script as shown below, running it DOES fix the warnings. ![image](https://user-images.githubusercontent.com/130684415/233176785-23901884-1160-4e78-8d23-7b7d08e8dcab.png) Would it be harmful to have the script look for ANY "${model}_host.db(.new)" OR "${model}_host_v7.db(.new)" files and patch all of them, regardless of DSM version installed?
Author
Owner

@007revad commented on GitHub (Apr 21, 2023):

@Greg-ATG

I'm working on changing the script to ignore the DSM version and update all the host db files it finds. This will avoid the issue you had, and cater for any changes by new DSM versions in future. As it requires quite a few changes it'll take a couple of days to finish and test.

I'd like to get to you to test it when I'm finished.

<!-- gh-comment-id:1517474595 --> @007revad commented on GitHub (Apr 21, 2023): @Greg-ATG I'm working on changing the script to ignore the DSM version and update all the host db files it finds. This will avoid the issue you had, and cater for any changes by new DSM versions in future. As it requires quite a few changes it'll take a couple of days to finish and test. I'd like to get to you to test it when I'm finished.
Author
Owner

@Greg-ATG commented on GitHub (Apr 21, 2023):

Sounds good, no rush. I'll help out however I can.

<!-- gh-comment-id:1518146424 --> @Greg-ATG commented on GitHub (Apr 21, 2023): Sounds good, no rush. I'll help out however I can.
Author
Owner

@007revad commented on GitHub (Apr 23, 2023):

@Greg-ATG

I've got a new pre-release version that should solve the issue of DSM 6 using X_v7.db files after downgrading from DSM 7 to DSM 6. https://github.com/007revad/Synology_HDD_db/releases/tag/v2.2.40

It also has a new --restore option to undo all changes previously made by the script so you can test v2.2.40 with clean db files and synoinfo.conf file.

Can you please:

  1. First run v2.2.40 with the --restore option: sudo syno_hdd_db --restore
  2. Then check that your drives show as unsupported.
  3. Then run it again with your preferred options and check that your drives do NOT show as unsupported.

It works perfectly in my test environment where I simulate having two expansion units and an M.2 card. But as I don't actually have an expansion unit or M.2 card I wouldn't be surprised if there's a bug that I didn't find.

<!-- gh-comment-id:1519013799 --> @007revad commented on GitHub (Apr 23, 2023): @Greg-ATG I've got a new pre-release version that should solve the issue of DSM 6 using X_v7.db files after downgrading from DSM 7 to DSM 6. https://github.com/007revad/Synology_HDD_db/releases/tag/v2.2.40 It also has a new --restore option to undo all changes previously made by the script so you can test v2.2.40 with clean db files and synoinfo.conf file. Can you please: 1. First run v2.2.40 with the --restore option: `sudo syno_hdd_db --restore` 2. Then check that your drives show as unsupported. 3. Then run it again with your preferred options and check that your drives do NOT show as unsupported. It works perfectly in my test environment where I simulate having two expansion units and an M.2 card. But as I don't actually have an expansion unit or M.2 card I wouldn't be surprised if there's a bug that I didn't find.
Author
Owner

@007revad commented on GitHub (Apr 23, 2023):

@Greg-ATG

Sorry, that pre-release version was mistakenly the same v2.1.38 !

Try the new https://github.com/007revad/Synology_HDD_db/releases/tag/v2.2.42

<!-- gh-comment-id:1519187824 --> @007revad commented on GitHub (Apr 23, 2023): @Greg-ATG Sorry, that pre-release version was mistakenly the same v2.1.38 ! Try the new https://github.com/007revad/Synology_HDD_db/releases/tag/v2.2.42
Author
Owner

@Greg-ATG commented on GitHub (Apr 24, 2023):

I ran the script with restore mode (nice feature!) with the following output:

image

However, the NAS reported NO unverified drives after a reboot. Upon closer inspection, I found the following:

  • rs2421rp+_host.db: Clean (no references to custom disk model)
  • rs2421rp+_host.db.new: Redundant references to custom disk model (see highlighting in screenshot below)
  • rs2421rp+_host_v7:db: Single reference to custom disk model
  • rs2421rp+_host_v7.db.new: Clean (in hindsight, this was likely not the case as I was only looking at the beginning of the file, not the end, where it seems this script seems to add it for v7 files)

I'm not sure if these inconsistencies are bugs with the restore mode or (more likely) a result of the tinkering I had previously done with these files.

image

I manually cleaned up all of these files and rebooted the NAS, which resulted in the expected unverified disks warning. I then ran the script with my preferred parameters (-showedits -noupdate -m2) and I'm happy to report that there was no unverified disk warning after rebooting and all the DB files look correct (it was at this point that I realized that this script adds the modifications to v7 files at the end, hence my note up above about v7.db.new, and I manually cleaned up the v7.db.new.bak file).

In the interest of science, I ran the script in restore mode again and can confirm that everything was restored correctly. I than ran and rebooted twice using my preferred parameters. After the first execution, everything looked good, but after the second one, it appears duplicate entries were added to the regular (v6) DB files:

image

This is confirmed by the script log (for the second execution), which shows that the v6 files were "re-patched" while the v7 files were correctly detected as "already exists":

image

FYI, the v7 DB files look fine after the second execution. Also, we don't have any expansion units or M.2 cards in our system.

<!-- gh-comment-id:1520394583 --> @Greg-ATG commented on GitHub (Apr 24, 2023): I ran the script with restore mode (nice feature!) with the following output: ![image](https://user-images.githubusercontent.com/130684415/234019620-31d12071-8708-46cc-afc8-10df8491bd1b.png) However, the NAS reported NO unverified drives after a reboot. Upon closer inspection, I found the following: - rs2421rp+_host.db: Clean (no references to custom disk model) - rs2421rp+_host.db.new: Redundant references to custom disk model (see highlighting in screenshot below) - rs2421rp+_host_v7:db: Single reference to custom disk model - rs2421rp+_host_v7.db.new: Clean (in hindsight, this was likely not the case as I was only looking at the beginning of the file, not the end, where it seems this script seems to add it for v7 files) I'm not sure if these inconsistencies are bugs with the restore mode or (more likely) a result of the tinkering I had previously done with these files. ![image](https://user-images.githubusercontent.com/130684415/234021376-6623fd11-8660-400f-a6da-5c038d81ff49.png) I manually cleaned up all of these files and rebooted the NAS, which resulted in the expected unverified disks warning. I then ran the script with my preferred parameters (-showedits -noupdate -m2) and I'm happy to report that there was no unverified disk warning after rebooting and all the DB files look correct (it was at this point that I realized that this script adds the modifications to v7 files _at the end_, hence my note up above about v7.db.new, and I manually cleaned up the v7.db.new.bak file). In the interest of science, I ran the script in restore mode again and can confirm that everything was restored correctly. I than ran and rebooted twice using my preferred parameters. After the first execution, everything looked good, but after the second one, it appears duplicate entries were added to the regular (v6) DB files: ![image](https://user-images.githubusercontent.com/130684415/234042042-11d8a110-4b5e-4704-9ccf-fa5737ae54b8.png) This is confirmed by the script log (for the second execution), which shows that the v6 files were "re-patched" while the v7 files were correctly detected as "already exists": ![image](https://user-images.githubusercontent.com/130684415/234043541-26741a02-8012-4295-a011-dbe5395fc247.png) FYI, the v7 DB files look fine after the second execution. Also, we don't have any expansion units or M.2 cards in our system.
Author
Owner

@007revad commented on GitHub (Apr 24, 2023):

Thank you for all your testing.

I chose the easy way to restore the files by restoring them from their backups. I did realise that there was a risk of restoring an edited file if the .db file was edited before it was backed up. I'll work out how to make DSM download clean db files.

For the duplicating entries in the v6 db files I've noticed the script is also creating an extra file named rs2421rp+_host.dbr with an "r" on the end. I haven't figured out what is causing this yet.

I've found the cause of the duplicated entries in the v6 db files which has existed since v1.3.32 but nobody (including me) noticed until now. It will be an easy fix.

<!-- gh-comment-id:1520802799 --> @007revad commented on GitHub (Apr 24, 2023): Thank you for all your testing. I chose the easy way to restore the files by restoring them from their backups. I did realise that there was a risk of restoring an edited file if the .db file was edited before it was backed up. I'll work out how to make DSM download clean db files. For the duplicating entries in the v6 db files I've noticed the script is also creating an extra file named rs2421rp+_host.db**r** with an "r" on the end. I haven't figured out what is causing this yet. I've found the cause of the duplicated entries in the v6 db files which has existed since v1.3.32 but nobody (including me) noticed until now. It will be an easy fix.
Author
Owner

@007revad commented on GitHub (Apr 24, 2023):

@Greg-ATG
I've fixed the multiplying entries in the v6 db files issue.
I've also fixed the v6 .db files being copied with a .dbr extension issue (which has existed since v1.1.10).

https://github.com/007revad/Synology_HDD_db/releases/tag/v2.2.43

<!-- gh-comment-id:1520918910 --> @007revad commented on GitHub (Apr 24, 2023): @Greg-ATG I've fixed the multiplying entries in the v6 db files issue. I've also fixed the v6 .db files being copied with a .dbr extension issue (which has existed since v1.1.10). https://github.com/007revad/Synology_HDD_db/releases/tag/v2.2.43
Author
Owner

@Greg-ATG commented on GitHub (Apr 25, 2023):

Everything seems to work great! No redundant entries and the .dbr/.newr files don't reappear anymore.

It would be pretty cool if you can pull clean DB copies from Synology when using "--restore"! If that doesn't work out, maybe it makes sense to create two types of backups:

  1. .origbak
  2. .lastbak

Where the "origbak" is created the first time the script runs (i.e. if "origbak" doesn't already exist) whereas any subsequent times the script runs and needs to make a change, it creates/overwrites "lastbak" so as not to overwrite the original backup. I suppose this would require two different restore modes (restore vs restorelast).

Anyway, I appreciate all your work in getting these issues resolved!

<!-- gh-comment-id:1522041138 --> @Greg-ATG commented on GitHub (Apr 25, 2023): Everything seems to work great! No redundant entries and the .dbr/.newr files don't reappear anymore. It would be pretty cool if you can pull clean DB copies from Synology when using "--restore"! If that doesn't work out, maybe it makes sense to create two types of backups: 1. <filename>.origbak 2. <filename>.lastbak Where the "origbak" is created the first time the script runs (i.e. if "origbak" doesn't already exist) whereas any subsequent times the script runs and needs to make a change, it creates/overwrites "lastbak" so as not to overwrite the original backup. I suppose this would require two different restore modes (restore vs restorelast). Anyway, I appreciate all your work in getting these issues resolved!
Author
Owner

@007revad commented on GitHub (Apr 26, 2023):

It would be pretty cool if you can pull clean DB copies from Synology when using "--restore"

Funny you should mention that. I added that to the develop branch, 3 hours before you commented :)

I just need to change one of the images in the readme to show the --restore option then I'll push it to the main branch and create a new release. Though I should probably edit the script to also delete any .dbr and .newr files.

<!-- gh-comment-id:1522671234 --> @007revad commented on GitHub (Apr 26, 2023): > It would be pretty cool if you can pull clean DB copies from Synology when using "--restore" Funny you should mention that. I added that to the develop branch, 3 hours before you commented :) I just need to change one of the images in the readme to show the --restore option then I'll push it to the main branch and create a new release. Though I should probably edit the script to also delete any .dbr and .newr files.
Author
Owner

@007revad commented on GitHub (Apr 26, 2023):

I forgot to mention that v2.2.44 is now available. https://github.com/007revad/Synology_HDD_db/releases/tag/v2.2.44

  • Added --restore info to --help
  • Updated restore option to download the latest db files from Synology
  • Now warns you if you try to run it in sh with "sh scriptname.sh"
<!-- gh-comment-id:1522793944 --> @007revad commented on GitHub (Apr 26, 2023): I forgot to mention that v2.2.44 is now available. https://github.com/007revad/Synology_HDD_db/releases/tag/v2.2.44 - Added --restore info to --help - Updated restore option to download the latest db files from Synology - Now warns you if you try to run it in sh with "sh scriptname.sh"
Author
Owner

@tmnext commented on GitHub (Apr 27, 2023):

I am having the same problem on DS1823xs+ on DSM7.2 RC1; HDDs work, but firmware warning. The --restore in v2.44 leads to:
rm: cannot remove '/var/lib/disk-compatibility/*dbr': No such file or directory
rm: cannot remove '/var/lib/disk-compatibility/*db.newr': No such file or directory

<!-- gh-comment-id:1524933260 --> @tmnext commented on GitHub (Apr 27, 2023): I am having the same problem on DS1823xs+ on DSM7.2 RC1; HDDs work, but firmware warning. The --restore in v2.44 leads to: rm: cannot remove '/var/lib/disk-compatibility/*dbr': No such file or directory rm: cannot remove '/var/lib/disk-compatibility/*db.newr': No such file or directory
Author
Owner

@007revad commented on GitHub (Apr 27, 2023):

Those messages are harmless, but I have now prevented them from appearing in v2.2.45

Is the first time you've used the script?

Does it find your drives?

Can you run the script with -s or --show then copy and paste the output from the script here, or paste a screenshot.

<!-- gh-comment-id:1524992497 --> @007revad commented on GitHub (Apr 27, 2023): Those messages are harmless, but I have now prevented them from appearing in v2.2.45 Is the first time you've used the script? Does it find your drives? Can you run the script with -s or --show then copy and paste the output from the script here, or paste a screenshot.
Author
Owner

@tmnext commented on GitHub (Apr 27, 2023):

Yes first time, the system is new. The 2.45 does run successfully, here is the output from one of the drives:

WD221KFGX-68B9KN0": {
"0A83": {
"compatibility_interval": [
{
"compatibility": "support",
"not_yet_rolling_status": "support",
"fw_dsm_update_status_notify": false,
"barebone_installable": true
}

All good, with the exception that the firmware warning does not go away (interestingly, it does not appear for the Samsung 970 M.2 drive, which runs with your enable_m2 script

<!-- gh-comment-id:1525020330 --> @tmnext commented on GitHub (Apr 27, 2023): Yes first time, the system is new. The 2.45 does run successfully, here is the output from one of the drives: WD221KFGX-68B9KN0": { "0A83": { "compatibility_interval": [ { "compatibility": "support", "not_yet_rolling_status": "support", "fw_dsm_update_status_notify": false, "barebone_installable": true } All good, with the exception that the firmware warning does not go away (interestingly, it does not appear for the Samsung 970 M.2 drive, which runs with your enable_m2 script
Author
Owner

@tmnext commented on GitHub (Apr 27, 2023):

More output:

Synology_HDD_db v2.2.45
DS1823xs+ DSM 7.2-64551 RC
Using options: -n

HDD/SSD models found: 3
SSD 860 EVO 4TB,2B6Q
WD221KFGX-68B9KN0,0A83
WUH721414ALE6L4,W240

M.2 drive models found: 2
Samsung SSD 960 PRO 2TB,ccc
Samsung SSD 970 EVO Plus 2TB,ccc

No M.2 cards found

No Expansion Units found

SSD 860 EVO 4TB already exists in ds1823xs+_host_v7.db
SSD 860 EVO 4TB already exists in ds1823xs+_host_v7.db.new
WD221KFGX-68B9KN0 already exists in ds1823xs+_host_v7.db
WD221KFGX-68B9KN0 already exists in ds1823xs+_host_v7.db.new
WUH721414ALE6L4 already exists in ds1823xs+_host_v7.db
WUH721414ALE6L4 already exists in ds1823xs+_host_v7.db.new
Samsung SSD 960 PRO 2TB already exists in ds1823xs+_host_v7.db
Samsung SSD 960 PRO 2TB already exists in ds1823xs+_host_v7.db.new
Samsung SSD 970 EVO Plus 2TB already exists in ds1823xs+_host_v7.db
Samsung SSD 970 EVO Plus 2TB already exists in ds1823xs+_host_v7.db.new

Support disk compatibility already enabled.

Support memory compatibility already enabled.

M.2 volume support already enabled.

Disabled drive db auto updates.

DSM successfully checked disk compatibility.

<!-- gh-comment-id:1525033381 --> @tmnext commented on GitHub (Apr 27, 2023): More output: Synology_HDD_db v2.2.45 DS1823xs+ DSM 7.2-64551 RC Using options: -n HDD/SSD models found: 3 SSD 860 EVO 4TB,2B6Q WD221KFGX-68B9KN0,0A83 WUH721414ALE6L4,W240 M.2 drive models found: 2 Samsung SSD 960 PRO 2TB,ccc Samsung SSD 970 EVO Plus 2TB,ccc No M.2 cards found No Expansion Units found SSD 860 EVO 4TB already exists in ds1823xs+_host_v7.db SSD 860 EVO 4TB already exists in ds1823xs+_host_v7.db.new WD221KFGX-68B9KN0 already exists in ds1823xs+_host_v7.db WD221KFGX-68B9KN0 already exists in ds1823xs+_host_v7.db.new WUH721414ALE6L4 already exists in ds1823xs+_host_v7.db WUH721414ALE6L4 already exists in ds1823xs+_host_v7.db.new Samsung SSD 960 PRO 2TB already exists in ds1823xs+_host_v7.db Samsung SSD 960 PRO 2TB already exists in ds1823xs+_host_v7.db.new Samsung SSD 970 EVO Plus 2TB already exists in ds1823xs+_host_v7.db Samsung SSD 970 EVO Plus 2TB already exists in ds1823xs+_host_v7.db.new Support disk compatibility already enabled. Support memory compatibility already enabled. M.2 volume support already enabled. Disabled drive db auto updates. DSM successfully checked disk compatibility.
Author
Owner

@007revad commented on GitHub (Apr 27, 2023):

If you've upgraded the memory with non-Synology memory you should include the -r option when running the script, to prevent these annoying notifications.

Try the -f option. So you'd want to use -nfr

The -f option should get rid of the firmware warnings until I can work out what 7.2 RC has changed since 7.2 beta.

<!-- gh-comment-id:1525078449 --> @007revad commented on GitHub (Apr 27, 2023): If you've upgraded the memory with non-Synology memory you should include the `-r` option when running the script, to prevent [these annoying notifications](https://github.com/007revad/Synology_HDD_db/blob/main/images/ram_warning.png). Try the `-f` option. So you'd want to use `-nfr` The -f option should get rid of the firmware warnings until I can work out what 7.2 RC has changed since 7.2 beta.
Author
Owner

@tmnext commented on GitHub (Apr 27, 2023):

Thanks for your prompt reply and the awesome work here :) I just tried -nfr with v2.2.45, but there is no change in output and warnings are still the same. No rush, I'll just wait till you get around to it. Thanks again.

<!-- gh-comment-id:1525103581 --> @tmnext commented on GitHub (Apr 27, 2023): Thanks for your prompt reply and the awesome work here :) I just tried -nfr with v2.2.45, but there is no change in output and warnings are still the same. No rush, I'll just wait till you get around to it. Thanks again.
Author
Owner

@tmnext commented on GitHub (Apr 27, 2023):

After restart (DSM7.2 RC1), the storage manager is not visible anymore. Not sure if thats related to 7.2 RC1 or the script :)

<!-- gh-comment-id:1525538470 --> @tmnext commented on GitHub (Apr 27, 2023): After restart (DSM7.2 RC1), the storage manager is not visible anymore. Not sure if thats related to 7.2 RC1 or the script :)
Author
Owner

@007revad commented on GitHub (Apr 27, 2023):

Had you run my Synology_enable_M2_volume or Synology_enable_Deduplication scripts? They caused a couple of minor issues in 7.2-64213 Beta - which Synology fixed in 7.2-64216 Beta.

Otherwise it sounds like a 7.2 RC1 bug.

You can run syno_hdd_db.sh --restore to undo the changes made by syno_hdd_db then check if storage manager is back.

<!-- gh-comment-id:1526685071 --> @007revad commented on GitHub (Apr 27, 2023): Had you run my [Synology_enable_M2_volume](https://github.com/007revad/Synology_enable_M2_volume) or [Synology_enable_Deduplication](https://github.com/007revad/Synology_enable_Deduplication) scripts? They caused a couple of [minor issues](https://github.com/007revad/Synology_enable_M2_volume/blob/main/known_issues.md) in 7.2-6421**3** Beta - which Synology fixed in 7.2-64216 Beta. Otherwise it sounds like a 7.2 RC1 bug. You can run `syno_hdd_db.sh --restore` to undo the changes made by syno_hdd_db then check if storage manager is back.
Author
Owner

@tmnext commented on GitHub (Apr 28, 2023):

I repaired it by taking out all disks, adding a fresh one in empty slot one and installing DSM7.2RC1, re-inserting the old disks same slots while the system is running, and they popped up as "online assembly". All good - procedure might be useful for others.

This is now slightly off-topic, but I am running your hdd_db and enable_m2 scripts. Why do I need the enable_dedupe as well?

<!-- gh-comment-id:1526951244 --> @tmnext commented on GitHub (Apr 28, 2023): I repaired it by taking out all disks, adding a fresh one in empty slot one and installing DSM7.2RC1, re-inserting the old disks same slots while the system is running, and they popped up as "online assembly". All good - procedure might be useful for others. This is now slightly off-topic, but I am running your hdd_db and enable_m2 scripts. Why do I need the enable_dedupe as well?
Author
Owner

@007revad commented on GitHub (Apr 28, 2023):

You don't need to run enable_dedupe as well. I was just wondering if you had used either enable_m2_volume OR enable_dedupe before storage manager went missing.

  • Both enable_m2_volume and enable_dedupe make the same change to the same file to make DSM think your drives are Synology drives.
  • The only difference between those 2 scripts is that enable_dedupe also enables deduplication in synoinfo.conf and it checks the NAS has enough memory for deduplication (32GB minimum).

Are you still getting the firmware warning? And have you rebooted yet to see if the problem returns?

<!-- gh-comment-id:1527005624 --> @007revad commented on GitHub (Apr 28, 2023): You don't need to run enable_dedupe as well. I was just wondering if you had used either enable_m2_volume OR enable_dedupe before storage manager went missing. - Both enable_m2_volume and enable_dedupe make the same change to the same file to make DSM think your drives are Synology drives. - The only difference between those 2 scripts is that enable_dedupe also enables deduplication in synoinfo.conf and it checks the NAS has enough memory for deduplication (32GB minimum). Are you still getting the firmware warning? And have you rebooted yet to see if the problem returns?
Author
Owner

@tmnext commented on GitHub (Apr 28, 2023):

<Are you still getting the firmware warning? And have you rebooted yet to see if the problem returns?>

After restart, system normal, but still showing firmware warnings for all non-M2 drives.

<!-- gh-comment-id:1527133168 --> @tmnext commented on GitHub (Apr 28, 2023): <Are you still getting the firmware warning? And have you rebooted yet to see if the problem returns?> After restart, system normal, but still showing firmware warnings for all non-M2 drives.
Author
Owner

@007revad commented on GitHub (Apr 28, 2023):

Someone with a DS918+ and 7.2 RC has also reported that lost access to Storage Manager, Snapshot Replication app, and most of the Hardware tab in control panel (fan speed, beep setting, etc). Also in Control Panel/Info Center "Model name" is listed as "Unknown Model" (instead of DS918+) and "Fan Speed Mode" is blank. It sounds like it may be because DSM doesn't know what the model is any more.

This tells me that the issue affects not just business models or recent models but probably all models.

I'm going to check all the Synology forums to see if it is a known RC bug. If it's not then I'll switch off my DS1821+, pull my HDDs, install a spare drive then run my scripts, one at a time, to see if I can reproduce the issue.

<!-- gh-comment-id:1528101895 --> @007revad commented on GitHub (Apr 28, 2023): Someone with a DS918+ and 7.2 RC has also reported that lost access to Storage Manager, Snapshot Replication app, and most of the Hardware tab in control panel (fan speed, beep setting, etc). Also in Control Panel/Info Center "Model name" is listed as "Unknown Model" (instead of DS918+) and "Fan Speed Mode" is blank. It sounds like it may be because DSM doesn't know what the model is any more. This tells me that the issue affects not just business models or recent models but probably all models. I'm going to check all the Synology forums to see if it is a known RC bug. If it's not then I'll switch off my DS1821+, pull my HDDs, install a spare drive then run my scripts, one at a time, to see if I can reproduce the issue.
Author
Owner

@007revad commented on GitHub (Apr 29, 2023):

I repaired it by taking out all disks, adding a fresh one in empty slot one and installing DSM7.2RC1, re-inserting the old disks same slots while the system is running, and they popped up as "online assembly".

When you wrote "and installing DSM 7.2 RC1" did you mean you created a storage pool on that drive so it had a clean copy of DSM 7.2 RC1 on it's system partition, or that you were able to install the DSM 7.2 RC1 .pat file again?

<!-- gh-comment-id:1528487385 --> @007revad commented on GitHub (Apr 29, 2023): > I repaired it by taking out all disks, adding a fresh one in empty slot one and installing DSM7.2RC1, re-inserting the old disks same slots while the system is running, and they popped up as "online assembly". When you wrote "and installing DSM 7.2 RC1" did you mean you created a storage pool on that drive so it had a clean copy of DSM 7.2 RC1 on it's system partition, or that you were able to install the DSM 7.2 RC1 .pat file again?
Author
Owner

@tmnext commented on GitHub (Apr 29, 2023):

To clarify: This was exactly what had happened to mine (didnt want to overload github here with off-topic things), model unknown, storage manager gone, half the tabs gone; system felt still snappy fast, all data intact, almost as if there was a "rescue" mini-DSM underneath.

As for your question: Pulled out all drives with system off, added a fresh one (never used before), let web-assistant install by uploading DSM7.2RC1 pat file via web-assistant, but did NOT create a new pool or volume when the system was up. Gave it the same name and same account name. Then during system being up, I inserted all my disk in the same slots (that only worked cause I didn't populate all slots). At that time the system offered all my three old pools back as "online assembly possible". (I also tested turning system back off before old hdd insert and then inserted old drives, but it seemed the system would then choose randomly a DSM partition from an old drive and same problem occurred, so that did not work; only old hdd inserting when system is running via fresh hdd worked).

To top it off, I also used "full volume encryption" when I created the volumes the first time around, and during online assembly I was nicely asked to re-create the key vault with my old password, and system offered to upload the keys for each volume via my laptop, all that without having a working volume yet. Only then I re-run your hdd_db script 2.45 and the enable m2_volume script; re-started several times to make sure and it seems to work fine since then. Maybe it had to do with me coming from beta to RC1 the first time around (not the beta 64123, one after that, I think it ended in 6), or that I used the -f function which I didnt the second time. Since the system is new, pretty sure there isnt anything else I did with the system, and no other scripts are being used. I also used the automatic update function of hdd_db, original script was a few days older.

<!-- gh-comment-id:1528615030 --> @tmnext commented on GitHub (Apr 29, 2023): To clarify: This was exactly what had happened to mine (didnt want to overload github here with off-topic things), model unknown, storage manager gone, half the tabs gone; system felt still snappy fast, all data intact, almost as if there was a "rescue" mini-DSM underneath. As for your question: Pulled out all drives with system off, added a fresh one (never used before), let web-assistant install by uploading DSM7.2RC1 pat file via web-assistant, but did NOT create a new pool or volume when the system was up. Gave it the same name and same account name. Then during system being up, I inserted all my disk in the same slots (that only worked cause I didn't populate all slots). At that time the system offered all my three old pools back as "online assembly possible". (I also tested turning system back off before old hdd insert and then inserted old drives, but it seemed the system would then choose randomly a DSM partition from an old drive and same problem occurred, so that did not work; only old hdd inserting when system is running via fresh hdd worked). To top it off, I also used "full volume encryption" when I created the volumes the first time around, and during online assembly I was nicely asked to re-create the key vault with my old password, and system offered to upload the keys for each volume via my laptop, all that without having a working volume yet. Only then I re-run your hdd_db script 2.45 and the enable m2_volume script; re-started several times to make sure and it seems to work fine since then. Maybe it had to do with me coming from beta to RC1 the first time around (not the beta 64123, one after that, I think it ended in 6), or that I used the -f function which I didnt the second time. Since the system is new, pretty sure there isnt anything else I did with the system, and no other scripts are being used. I also used the automatic update function of hdd_db, original script was a few days older.
Author
Owner

@007revad commented on GitHub (Apr 30, 2023):

I've noticed that the .db files in 7.2 RC include "smart_test_ignore" and "smart_attr_ignore" for all drives. In 7.2 beta and 7.1 it was only Synology drives that had those.

Synology drives have:
"smart_test_ignore": true,
"smart_attr_ignore": true

And in 7.2 RC 3rd party drives now have:
"smart_test_ignore": false,
"smart_attr_ignore": false

While manually adding your drives to a clean DS1832+ DSM 7.2 RC host db file (for you to test) I noticed that the firmware versions don't look correct.

  • The SSD 860 EVO 4TB firmware version I'd expect to be like "RVT02B6Q" and not just "2B6Q".
  • The WD firmware version I'd expect to be like "80.00A83" and not just "0A83".
  • I'm not sure what to expect for the Ultrastar but maybe like "LDGNW240" and not just "W240".
  • I'm surprised that the NVMe drives' firmware version is "ccc". Maybe ccc is returned when no firmware version is available.

Can you run the following commands and let me know what they return:

for d in /sys/block/sata*; do echo $(cat "$d/device/rev"); done

cat /sys/block/nvme0n1/device/firmware_rev

cat /sys/block/nvme1n1/device/firmware_rev

If the last 2 commands return "No such file or directory" then run these commands:

cat /sys/block/nvme0n1/device/rev

cat /sys/block/nvme1n1/device/rev

<!-- gh-comment-id:1528911436 --> @007revad commented on GitHub (Apr 30, 2023): I've noticed that the .db files in 7.2 RC include "smart_test_ignore" and "smart_attr_ignore" for all drives. In 7.2 beta and 7.1 it was only Synology drives that had those. Synology drives have: "smart_test_ignore": true, "smart_attr_ignore": true And in 7.2 RC 3rd party drives now have: "smart_test_ignore": false, "smart_attr_ignore": false While manually adding your drives to a clean DS1832+ DSM 7.2 RC host db file (for you to test) I noticed that the firmware versions don't look correct. - The SSD 860 EVO 4TB firmware version I'd expect to be like "RVT02B6Q" and not just "2B6Q". - The WD firmware version I'd expect to be like "80.00A83" and not just "0A83". - I'm not sure what to expect for the Ultrastar but maybe like "LDGNW240" and not just "W240". - I'm surprised that the NVMe drives' firmware version is "ccc". Maybe ccc is returned when no firmware version is available. Can you run the following commands and let me know what they return: `for d in /sys/block/sata*; do echo $(cat "$d/device/rev"); done` `cat /sys/block/nvme0n1/device/firmware_rev` `cat /sys/block/nvme1n1/device/firmware_rev` If the last 2 commands return "No such file or directory" then run these commands: `cat /sys/block/nvme0n1/device/rev` `cat /sys/block/nvme1n1/device/rev`
Author
Owner

@007revad commented on GitHub (May 5, 2023):

@70m7E not responding

<!-- gh-comment-id:1535633326 --> @007revad commented on GitHub (May 5, 2023): @70m7E not responding
Sign in to join this conversation.
No milestone
No project
No assignees
1 participant
Notifications
Due date
The due date is invalid or out of range. Please use the format "yyyy-mm-dd".

No due date set.

Dependencies

No dependencies set.

Reference
starred/Synology_HDD_db#12
No description provided.