NSS Shared DB: Difference between revisions
(partial editorial pass - some wordsmithing - more to be done) |
|||
Line 765: | Line 765: | ||
I think option 3 probably provides the best solution for all worlds. | I think option 3 probably provides the best solution for all worlds. | ||
===== Database Merge ===== | |||
==== Layering ==== | ==== Layering ==== |
Revision as of 23:50, 25 January 2008
NSS has been using an old version of the Berkeley DataBase as its database engine since Netscape Navigator 2.0 in 1994. That version of the database engine is commonly described in NSS documents as "DBM". That engine has a number of limitations. One of the most severe limitations concerns the number of processes that may share a database file. While any process has a DBM file open for writing, NO other process may access it in any way. Multiple processes may share a DBM database ONLY if they ALL access it READ-ONLY. Processes cannot share a DBM database file if ANY of them wants to update it.
This limitation has been cumbersome for applications that wish to use NSS. Applications that want to share databases have resorted to these strategies:
- Synchronized updates, with application down time: The applications share the database read-only. If any update is desired, all the applications are shut down, and a database update program performs the update, then all the applications are restarted in read-only mode. Some server products, for example, have an administration program that stops the servers, updates the database that they share, and then restarts the servers. This results in undesirable downtime and desired database changes are delayed until the next interval in which such downtime is acceptable.
- Multiple copies with duplicated updates. Each application keeps its own copy of its databases, and applications communicate their changes to each other, so that each application may apply received changes to its own DB. FireFox and Thunderbird are examples of this. When one of those applications gets a new certificate and private key, the user may "export" that pair to a PKCS#12 file, and then import that file into the other application. Most users never master these steps, and so have databases entirely out of sync.
These workarounds for the DBM engine's limitations are sufficiently onerous that they prevent many applications from adopting NSS. The desire to make NSS more ubiquitous now motivates the elimination of these limitations. There is a strong desire to make NSS be the native OS network security service for Linux.
In 2001 NSS was modified to enable applications to supply their own database engines. Applications could share a common database if they supplied their own shared database implementation, and configured NSS to use it.
Today, there exists a process level, ACID, open source, and widely available database engine with which multiple processes may simultaneous have read and write access to a shared database. It is named SQLite. The NSS team proposes to leverage this database to give all NSS-based applications Shared Database access.
Besides the inability to share databases, other issues with NSS DBM database scheme include:
- A more flexible schema which can store meta information about certificates and keys, such as finer grained information about trust for certificates.
- The need to match the underlying certificate and key storage with its reflection into NSS (that is, PKCS #11).
- To satisfy the FIPS requirements for integrity checks on keys and trust objects, NSS must be able to store integrity information in the databases.
Where we are today
At initialization time, the application gives NSS a string that it uses as the pathname of a directory to store NSS's security and configuration data. NSS typically stores 3 dbm files in that directory:
- cert8.db - stores publicly accessible objects (certs, CRLs, S/MIME records).
- key3.db - stores the private keys.
- secmod.db - stores the PKCS #11 module configuration.
Also in that directory:
- If it has very large security objects (such as large CRLs), NSS will store them in files in a subdirectory named cert8.dir.
- If the cert8.db and/or key3.db files are missing, NSS will read data from older versions of those databases (e.g., cert7.db, cert5.db, if they exist)
and may build new cert8.db and/or key3.db files.
These files are all accessed exclusively by the softoken shared library, making it the only NSS library that must be linked with libdbm.
The application-supplied database feature
If the initialization string given to NSS starts with 'multiaccess:', NSS does not use it as a directory pathname. Instead, NSS parses the string as follows:
multiaccess:appName[:directory]
Where:
multiaccess is a keyword.
appName uniquely identifies a group of applications which share an
application-supplied database, effectively a new database name.
directory is the pathname for a directory containing NSS DBM databases
whose contents will be used to update the application-supplied database during NSS initialization.
In the presence of a multiaccess initialization string, during initialization NSS will try to find a shared database named librdb.so (rdb.dll on Windows) in its path and load it. This shared library is expected to implement a superset of the old dbm interface. The main entry point is rdbopen, which will be passed the appName, database name, and open flags. The rdb shared library will pick a location or method to store the database (it may not necessarily be a file), then handle the raw db records from NSS. The records passed to and from this library use exactly the same schema and record formats as the records in the DBM library.
The proposal
We propose to replace key3.db and cert8.db with new SQL databases called key4.db and cert9.db. These new databases will store PKCS #11 token objects, the same types of objects whose contents are currently stored in cert8.db and key3.db
Optionally the new databases could be combined into a single database, cert9.db, where private and public objects are stored in separate tables. Softoken would automatically identify any cert9.db that also has an embedded key store and open that key store up, instead of opening a separate key4.db and cert9.db. However, in the first release, there will be no way to create a cert9.db containing both cert and key tables.
The new databases will be called 'shareable' databases. They may or may not be shared by multiple processes, but they are all capable of being shared.
schema
The schema for the database will be simple.
- Each row will represent an object.
- The row schema will contain the the Object ID and the list of known Attribute Types.
- Newer versions of NSS may add new attribute types on the fly as necessary (extending the schema).
- The Attribute values will be stored as binary blobs.
- Attributes that represent CK_ULONG values will be stored as 32-bit values in network byte order.
- For all other attributes, byte order is already specified by PKCS #11.
- Private attributes will be encrypted with a PKCS #5 PBE in the same way the pkcs8 private and secret key data is encrypted today.
- Softoken will only set those attributes appropriate for the given object. If The attribute is not appropriate it will be left blank. (Note that sqlite does not distinguish between a NULL attribute and an empty one. This will be handled by storing a special value which means 'NULL' when writing a NULL record.
- integrity will be maintained by a PBE based MAC on critical attributes.
Other data that is necessary for the proper operation of softoken, but that is not defined as part of any PKCS#11 objects, (such as data used to verify token user passwords), will be stored in separate tables in the database with their own schemas.
Database extension will be accomplished in 2 ways:
- New attribute types can augment the already-implemented attribute types for objects already implemented in softoken. Attributes of these new types can be added to older database objects, which will be detected because they will have 'invalid' values for these attributes. For example, we could add a new attribute type to hold additional extensions for certificate objects.
- Define new PKCS #11 object types. For example, we could add new objects to store mappings between various certificates pairs of cipher suites and host names.
Softoken will be able to store the following objects and attributes. In the table below, attributes marked CK_ULONG will be written to the database as a 32-bit network byte order unsigned integers. Attributes marked 'encrypted' will be encrypted with the token's pbe key, and attributes marked 'MACed' will be MACed with the token's PBE key.
Legal Attributes and objects
While the key and certificate database format is extensible, the initial implementation has to understand a particular subset of attributes. The following list of attribute types will be understood, and any special coding conditions for those attribute types.
- Stored in the key database:
- CKO_PRIVATE_KEY
- CKA_CLASS - CK_ULONG
- CKA_TOKEN
- CKA_PRIVATE
- CKA_LABEL
- CKA_MODIFIABLE
- CKA_ID
- CKA_START_DATE
- CKA_END_DATE
- CKA_DERIVE
- CKA_LOCAL
- CKA_KEY_TYPE - CK_ULONG
- CKA_DECRYPT
- CKA_SIGN
- CKA_SIGN_RECOVER
- CKA_UNWRAP
- CKA_SUBJECT
- CKA_SENSITIVE,
- CKA_EXTRACTABLE
- CKA_NSS_DB - (allowed, but not required)
- CKA_MODULUS - RSA only - MACed
- CKA_PUBLIC_EXPONENT - RSA only - MACed
- CKA_PRIVATE_EXPONENT - RSA only - encrypted
- CKA_PRIME_1 - RSA only - encrypted
- CKA_PRIME_2 - RSA only - encrypted
- CKA_EXPONENT_1 - RSA only - encrypted
- CKA_EXPONENT_2 - RSA only - encrypted
- CKA_COEFFICIENT - RSA only - encrypted
- CKA_SUBPRIME - DSA only - MACed
- CKA_PRIME - DH, DSA only - MACed
- CKA_BASE - DH, DSA only - MACed
- CKA_VALUE - DH, DSA, and ECC only - encrypted
- CKA_EC_PARAMS - ECC only - MACed
- CKO_SECRET_KEY
- CKA_CLASS - CK_ULONG
- CKA_TOKEN
- CKA_PRIVATE
- CKA_LABEL
- CKA_MODIFIABLE
- CKA_ID
- CKA_START_DATE
- CKA_END_DATE
- CKA_DERIVE
- CKA_LOCAL
- CKA_KEY_TYPE - CK_ULONG
- CKA_SENSITIVE
- CKA_EXTRACTABLE
- CKA_ENCRYPT
- CKA_DECRYPT
- CKA_SIGN
- CKA_VERIFY
- CKA_WRAP
- CKA_UNWRAP
- CKA_VALUE - encrypted
- CKO_PRIVATE_KEY
- Stored in the cert database:
- CKO_PUBLIC_KEY
- CKA_CLASS - CK_ULONG
- CKA_TOKEN
- CKA_PRIVATE
- CKA_LABEL
- CKA_MODIFIABLE
- CKA_ID
- CKA_START_DATE
- CKA_END_DATE
- CKA_DERIVE
- CKA_LOCAL
- CKA_KEY_TYPE - CK_ULONG
- CKA_ENCRYPT
- CKA_VERIFY
- CKA_VERIFY_RECOVER
- CKA_WRAP
- CKA_SUBJECT
- CKA_MODULUS - RSA only - MACed
- CKA_PUBLIC_EXPONENT - RSA only - MACed
- CKA_SUBPRIME - DSA only - MACed
- CKA_PRIME - DH, DSA only - MACed
- CKA_BASE - DH, DSA only - MACed
- CKA_VALUE - DH, DSA, ECC only - MACed
- CKA_EC_PARAMS - ECC only - MACed
- CKO_CERTIFICATE
- CKA_CLASS - CK_ULONG
- CKA_TOKEN
- CKA_PRIVATE
- CKA_LABEL
- CKA_MODIFIABLE
- CKA_CERTIFICATE_TYPE - CK_ULONG
- CKA_VALUE
- CKA_SUBJECT
- CKA_ISSUER
- CKA_SERIAL_NUMBER
- CKA_NSS_OVERRIDE_EXTENSIONS (sql database only) - MACed
- CKO_NSS_TRUST
- CKA_CLASS - CK_ULONG
- CKA_TOKEN
- CKA_PRIVATE
- CKA_LABEL
- CKA_MODIFIABLE
- CKA_ISSUER - MACed
- CKA_SERIAL_NUMBER- MACed
- CKA_CERT_SHA1_HASH- MACed
- CKA_CERT_MD5_HASH - MACed
- CKA_TRUST_SERVER_AUTH - CK_ULONG - MACed
- CKA_TRUST_CLIENT_AUTH - CK_ULONG - MACed
- CKA_TRUST_EMAIL_PROTECTION - CK_ULONG - MACed
- CKA_TRUST_CODE_SIGNING - CK_ULONG - MACed
- CKA_TRUST_STEP_UP_APPROVED - MACed
- CKO_NSS_CRL
- CKA_CLASS - CK_ULONG
- CKA_TOKEN
- CKA_PRIVATE
- CKA_LABEL
- CKA_MODIFIABLE
- CKA_SUBJECT
- CKA_VALUE
- CKA_NETSCAPE_URL
- CKA_NETSCAPE_KRL
- CKO_NSS_SMIME
- CKA_CLASS - CK_ULONG
- CKA_TOKEN
- CKA_PRIVATE
- CKA_LABEL
- CKA_MODIFIABLE
- CKA_SUBJECT
- CKA_NETSCAPE_EMAIL
- CKA_NETSCAPE_SMIME_TIMESTAMP
- CKA_VALUE
- CKO_PUBLIC_KEY
Special coding for CK_ULONG
All CK_ULONG values are encoded as 4 byte values, most significant byte first.
Special coding for encrypted entries
Encrypted entries are stored in the database as PKCS5 encoded blobs.
SEQUENCE { AlgorithmID algorithm, OctetString encryptedData };
The algorithm parameter must be a valid PKCS5 or PKCS12 PBE oid, and contain the appropriate salt value. NSS understands the following PBE oids:
- SEC_OID_PKCS12_PBE_WITH_SHA1_AND_TRIPLE_DES_CBC
- SEC_OID_PKCS12_V2_PBE_WITH_SHA1_AND_2KEY_TRIPLE_DES_CBC
- SEC_OID_PKCS12_V2_PBE_WITH_SHA1_AND_3KEY_TRIPLE_DES_CBC
- SEC_OID_PKCS5_PBE_WITH_MD2_AND_DES_CBC
- SEC_OID_PKCS5_PBE_WITH_MD5_AND_DES_CBC
- SEC_OID_PKCS5_PBE_WITH_SHA1_AND_DES_CBC
- SEC_OID_PKCS12_V2_PBE_WITH_SHA1_AND_128_BIT_RC2_CBC
- SEC_OID_PKCS12_V2_PBE_WITH_SHA1_AND_40_BIT_RC2_CBC
- SEC_OID_PKCS12_PBE_WITH_SHA1_AND_128_BIT_RC2_CBC
- SEC_OID_PKCS12_PBE_WITH_SHA1_AND_40_BIT_RC2_CBC
- SEC_OID_PKCS12_PBE_WITH_SHA1_AND_128_BIT_RC4
- SEC_OID_PKCS12_V2_PBE_WITH_SHA1_AND_2KEY_TRIPLE_DES_CBC
- SEC_OID_PKCS12_V2_PBE_WITH_SHA1_AND_3KEY_TRIPLE_DES_CBC
- SEC_OID_PKCS12_PBE_WITH_SHA1_AND_TRIPLE_DES_CBC
- SEC_OID_PKCS5_PBE_WITH_MD2_AND_DES_CBC
- SEC_OID_PKCS5_PBE_WITH_MD5_AND_DES_CBC
- SEC_OID_PKCS5_PBE_WITH_SHA1_AND_DES_CBC:
- SEC_OID_PKCS12_V2_PBE_WITH_SHA1_AND_128_BIT_RC2_CBC:
- SEC_OID_PKCS12_V2_PBE_WITH_SHA1_AND_40_BIT_RC2_CBC:
- SEC_OID_PKCS12_PBE_WITH_SHA1_AND_128_BIT_RC2_CBC:
- SEC_OID_PKCS12_PBE_WITH_SHA1_AND_40_BIT_RC2_CBC:
- SEC_OID_PKCS12_PBE_WITH_SHA1_AND_128_BIT_RC4:
- SEC_OID_PKCS5_PBES2 (actual encryption and prf algorithms are stored in parameters)
- Valid Encryption Algorithms
- SEC_OID_AES_128_CBC
- SEC_OID_AES_192_CBC
- SEC_OID_AES_256_CBC
- Valid PRF Algorithms
- SEC_OID_HMAC_SHA_1
- SEC_OID_HMAC_SHA256
- SEC_OID_HMAC_SHA384
- SEC_OID_HMAC_SHA512
- Valid Encryption Algorithms
- SEC_OID_PKCS5_PBMAC1 (actual hmac and prf algorthms are stored in parameters)
- Valid HMAC and PRF Algorithms
- SEC_OID_HMAC_SHA_1
- SEC_OID_HMAC_SHA256
- SEC_OID_HMAC_SHA384
- SEC_OID_HMAC_SHA512
- Valid HMAC and PRF Algorithms
The base key used in the PBE is the token password hashed with the token's global salt stored in the password entry. <In FIPS mode this resulting key is further transformed by setting it to value 'x' and using x to raise generator 'g' to the 'x' power modulo 'p'. The final result is the token key. g and p are stored with the base user's protection in the filesystem in the same directory as the certificate and key database. g and p should be considered 'secret' values.>*
- Purpose: by choosing an appropriate g and p, we can control password attempt rates from multiple processes. By making g and p secret, we can remove our reliance on the security of the PBE - the security of the password is at least as strong as the security of g and p.
This will be implemented by checking for the g/p file in the database directory. If no g/q file exists, key processing would continue as normal. If g/p does exist it will be used in the key formation. If a new key is being created, then g/p is create/used if in FIPS mode, and not create/used if not in FIPS mode. This way passwords continue to work even after mode switching.
Special coding for MACed entries
Some specialized entries in the database are MACed to either for FIPS protection or to protect trust integrity. MACed entries are only processed if the user has logged into the token. If the user has not logged into the token, then the MAC is not used.
MACs are not processed for the legacy database.
Entries which require MACs require that the user be logged into the token to write or modify the entry. On readback, they are not checked unless the user is logged in.
Tokens which have no key database (and therefore no master password) do not have any stored MACs.
Attributes that are MACed are:
- Trust object hashes and trust values
- public key values.
Certs themselves are considered properly authenticated by virtue of their signature, or their matching hash with the trust object.
MACS are formed by concatenating the object id (encoded as a database ULONG) with the attribute type (also encoded as a ULONG) and with the atttribute data. This concatenation is fed into a SHA1 HMAC keyed with the token key described in the 'Special coding for encrypted entries'. The resulting Mac is stored in a special table in the key database called the 'metadata'. This table could store other types of meta data as well in the future. Currently the password check is also stored in the metadata table.
Integrity entries are indexed by the string
sig_[cert/key]_{ObjectID}_{Attribute}
Where cert/key indicates which logical database the actual object is stored in, {ObjectID} is the id of the object and {Attribute} is the attribute this integrity entry is for.
Entries are check when the user tries to get and attribute. Once the attributes are fetched, GetAttributes loops through the attributes looking for attributes that require integrity checks. When such an attribute is found, the integrity check entry is fetched and an HMAC is calculated for the attribute. If the attribute matches the integrity entry, processing proceeds. If it doesn't that attribute data is cleared and CKR_SIGNATURE_INVALID is returned.
MACS are PBMAC1 data structures defined in pkcs5 2.0. Since pkcs5 v1 does not have integrity checks and pkcs12 has not definition for storing purely mac data,the shared DB integrity checks uses pkcs5 v2 to store that pbe and mac data.
Database coherency issues
In our previous database, we had issues with database corruption resulting in hard to diagnose issues. In order to mitigate that, This section analyses how various forms of corruption can affect the new database design, and possible ways of repairing that corruption in the field.
Hidden Meta-data records
The most obvious corruption issues surround data records which do not have direct visibility to the application, or NSS itself (beyond softoken). In the previous design these included such records as the SubjectList records which kept track of all the certs with a given subject. In this design, we have the following meta-data records:
- g/p files
- password entries
- MAC data
g/p files only exist in FIPs mode. Loss of g/p files are unrecoverable. All private data and MAC data will become inaccessible. This means all private and secret keys will be lost. Lost MAC data would have to be regenerated. This would require a special tool which modified the sqlite database directly.
loss of password entries would be almost as devistating. If there are any keys or MAC data, it is theoretically possible to verify a password using one of those entries. From that password the password entry could be regenerated. This would require a speical tool that modified the sqlite database directly.
loss of MAC data would interfere with the ability of the database to validate certificates. NOTE: this would only affect the databases ability to validate certificates while the user is logged in. Lost MAC data would have to be regenerated by walking the database and creating fresh MACs.
Linked records
Since each record represents a PKCS #11 object, the current database design is significantly less reliant on two related records being self-consistant. There are 2 areas, however where there is linkages: CKA_IDs and CKO_NSS_SMIME.
CKA_IDs link certificates with their related private keys. This is a PKCS #11 specified linkage. CKA_IDs are generated by NSS when keys are created. NSS sets this value to a hash related to the public key of the key pair. If this value is corrupted in the private key, then NSS will no longer be able to find that private key. A tool at the PKCS #11 level could repair such damage. If the CKA_ID of a certificate is corrupted, NSS will stop recognizing the certificate as a user certificate, and it will be unable to find the key associated with it. This can also be repaired by a PKCS #11 level tool. The corrupted CKA_ID also be repaired by deleting and reimporting the certificate. Both cases could be repaired by reimporting a pkcs12 file with the certificate and key.
If the Private key is deleted, but the corresponding public key is not, then NSS may be confused and think that the private key exists for the certificate. Also, if the public key is deleted, but the private key is not, NSS may be confused and think the private key does not exist for the certificate. Both of these cases will act correctly if the token is logged in. This kind of corruption can be repaired by reimporting the pkcs12 file, or by a PKCS #11 level tool to delete or restore the public key.
CKO_NSS_SMIME object holds the email address, subject of the S/MIME certificate, and the s/mime profile. In the legacy database, multiple email addresses could hold the same profile data and certificate subject, but only one s/mime profile data and subject would be allowed for each email address. In the new database, multiple independent email records can exist for the same email address. While S/MIME will function without CKO_NSS_SMIME objects, Certificates that verify multiple email address can not be found by the email addresses (other then the 'primary' address) without CKO_NSS_SMIME objects. If S/MIME records are corrupted, the Cerificates will not be findable for other email addresses. Unlike the legacy database, these records are destroyed by S/MIME records for other certificates with the same email address. Now if you have multiple certificates with the same email address, all those certificates can be found. This corruption can be repaired with a PKCS #11 level tool.
In order to maintain binary compatibility, the following keywords will be understood and used by softoken.
multiaccess:appName[:directory] works as it does today, including using the cert8/key3 record version.
dbm:directory opens an existing non-shared libdbm version 8 database.
sql:directory1[:directory2] opens a shared database,
cert9.db (& key4.db) in directory1 if cert9.db does exist. If the database does not exist, then directory2 is searched for a libdbm cert8.db and key3.db. If directory2 is not supplied, directory1 is searched.
extern:directory open a sql-like database by loading an external module, a. la. rdb and multiaccess:. This option would not be implemented in the initial release, but the extern: keyword would be reserved for future use.
Plain directory spec. For binary compatibility, the plain directory spec as the same as dbm:directory unless overridden with the NSS_DEFAULT_DATABASE environment variable. Applications will not need to change for this release of NSS. (particularly unfriendly applications that want to tweak with the actual database file). Users can force older applications to share the database with the environment variable. The environment variable only affects non-tagged directories.
When accessing the dbm: and multiaccess: directories, external shared library will be loaded which knows how to handle these legacy databases. This allows us to move much of the current mapping code into this shared library.
Secmod.db
In the dbm: and multiaccess: cases, there will be no changes to secmod.db.
In the sql: case, a new directory with separate flat files containing text files of the format specified in the PKCS #11 working group, but not yet included in any spec. This directory will be opened, locked, used, then closed (much like the current secmod.db). The directory will live as a sub directory of the directory that holds cert9.db/key4.db. As a directory of flat files it would not use the sqlite database to access these records. The file name should become pkcs11.txt.
User App Initialization and System App Initialization
One of the goals of making a shareable database version of NSS is to create a 'system crypto library' in which applications will automatically share database and configuration settings. In order for this to work, applications need to be able to open NSS databases from standard locations.
This design assumes that new NSS init functions will be defined for applications wanting to do 'standard user initialization', rather than building special knowledge into softoken or the database model. Note: This is different from the 2001 design, or and earlier prototype shared database, where the database code knew the location of the shared database.
Database Upgrade
To understand the issues of migration to the Shareable Database version of NSS from the traditional (legacy) versions, we group applications that use the new version of NSS into three 'modes' of operation, and into two types, of which there are five valid combinations.
Mode 1: Legacy applications which formerly used DBM databases, that upgrade to the new version of NSS without making any changes to the applications' code.
These applications will continue to use the legacy database support and the old database format. The applications cannot take advantage of new features in the shared database. In this Mode, the nssdbm3 shared library must be present. No update from legacy DBM to sharable is needed in this mode.
Summary: nssdbm3 shared library: Must be Present Update: none needed. Application Changes: none.
Mode 2: Applications that use the new shareable database engine, but choose not to share copies of their cert and key stores. They wish to keep separate copies of their databases. They may or may not have existing legacy DBM databases from older versions of those applications. (Some servers might be like this.)
These applications use the new shareable database engine. The first time the new application version runs, when NSS first creates the shareable databases, NSS will automatically detect instances of legacy databases and will upgrade those legacy databases to the new shareable ones without user interaction.
Summary: nssdbm3 shared library: need only for update Update: automatic, no application/user intervention. Application Changes: minor changes to Init, Changes to code that updates trust values or imports certificates with trust.
Mode 3: Applications that intend to share their keys and certs with other applications (the common case - browsers, mail clients, secure shells, vpns, etc.)
To achieve that sharing, these applications must share a single common set of databases. If older versions of these applications created legacy DBM databases, those legacy databases must be merged. To perform such a merge, NSS will need some extra support from the application, and possibly user intervention as well.
Summary: nssdbm3 shared library: need only for update Update: partially automated, requires application interaction, possibly user interaction. Application Changes: minor changes to Init, Changes to code that updates trust values or imports certificates with trust. Changes to aid in update
Type A: Type A applications are new versions of applications that existed before NSS supported sharable databases. They have existing legacy NSS databases. The new versions of these applications have been upgraded to the new NSS that supports shareable databases. If they intend to share the contents of those old databases, they need to merge the old database contents into the new ones.
All Mode 1 applications are type A and need nssdbm3 at all times. Mode 2 and Mode 3 applications of Type A need nssdbm3 to upgrade from the old legacy DBM databases to the new shareable databases. Mode 3 Type A applications need libnssdbm3 to merge data from legacy DBM databases into shareable ones. Mode 2 and Mode 3 type A applications do not need nssdbm3 except for upgrading and merging data from old legacy databases.
Type B: Type B applications are applications that never used any old version of NSS that supported only DBM databases. All NSS databases used by Type B applications are sharable databases. There are no legacy DBM databases for Type B applications. All Type B applications are either Mode 2 or Mode 3. Type B applications do no database upgrades, and do not need nssdbm3.
Upgrade complications
Upgrade complications only affect Type A applications. In order to merge a legacy database into an existing shareable database, NSS needs the password for both databases.
In Mode 1, NSS never needs to do an update or a merge.
Flow chart of NSS update actions for Mode 3: NSS initialize | V open legacy DB | V done
In Mode 2, the new database is uninitialized, so NSS only needs the password for the legacy database so it can read the secret keys in that legacy database, and so the new shared database password matches the old one. NSS can find the legacy database because it's in the same directory that the shared database lives in. NSS opens both databases at initialization time and uses the legacy database until the user authenticates (providing the legacy database password). NSS then uses that password to update the new shared database with the records from the old. The new database takes on the password from the legacy database, and the legacy database is closed. Future NSS initializations only open the new shared database. If the user never supplies a password, NSS will continue to treat the new shared database as uninitialized and will attempt to update from the old database on future opens until the update succeeds.
Flow chart of NSS update actions for Mode 2: NSS initialize | V open shared DB | V < is open shared DB > yes < initialized? >-------> done | no V < does legacy DB > no < exist? >-------> done | yes V open legacy DB | V no < does legacy DB > +-------< have a password? > | | yes | V | use legacy DB | until password | is supplied | | | V | < is password > no | < supplied? >-------> done | | +--------------+ V update shared DB | V close legacy DB | V done
In Mode 3, the new database may or may not be initialized. For the first mode 3 application, the new database will be uninitialized. NSS can proceed the with the same procedure as Mode 2. When the second and subsequent applications start, the new shared database will already be initialized with it's own password. We potentially need both passwords, the first to read the keys out of the legacy database, and the second to write those keys, as well as the required authentication values for any trust data. In order to preserve the existing data in the second application, NSS must be able to merge the data in the second application with the data in the existing database. We only want the merge to happen once, not every time the application starts. In Mode three we need to be able to identify when a database has already been updated, so the applications needs to tell us some unique identifier for its database. The application must be able to tell us where the old database lives, since it's a application private directory compared the the multiple application shared directory that the shared DB lives in.
Flow chart of NSS update actions for Mode 3: Start | V open shared DB | V < is open shared DB > yes < updated with given >-------> done < legacy DB? > | no V < does legacy DB > no < exist? >-----------+ | yes | V V open legacy DB mark given | legacy Updated---> done V no < does legacy DB > +-------< have a password? > | | yes | V | no < does legacy DB > +-------< have any private > | < or secret keys? > | | yes | V | use legacy DB | until password | is supplied | | | V | < is password > no | < supplied? >-------> exception A | | +--------------+ V no < does shared DB > +-------< have a password? > | | yes | V | yes < does shared DB's > +--------< password match > | < legacy DB's PW? > | | no | V | get shared DB | password | | | V | < is password > no | < supplied? >-------> exception B | | +--------------+ V update/merge shared DB | V close legacy DB | V done
exception A. Application needs to decide what happens if the legacy password is not supplied. Application can choose to:
- continue to use the legacy DB and try to update later.
- force NSS to mark legacy DB to be updated without actually updating the legacy DB (throwing away everything in the legacy DB).
- force NSS to update those objects it can from the legacy DB, throwing
away private keys and saved passwords.
exception B. Applications needs to decide what happens if the new shared DB password is not supplied. Application can choose to:
- continue to use the legacy DB and try to update later.
- force NSS to mark legacy DB to be updated without actually updating the legacy DB (throwing away everything in the legacy DB).
- force NSS to update those objects it can from the legacy DB,throwing away private keys and saved passwords, and trust information from the legacy DB.
- force NSS to reset the shared database password, throwing away private keys and saved passwords, and trust information rom the shared DB.
NOTE: Since we are potentially dealing with 2 different passwords, The application needs to be clear to the user which password it needs.
Merge Conflicts (Mode 3A only)
When merging databases in, it's possible (even likely), that the shared database and legacy DB's have the same objects. In the case of certs and keys, the merge is a simple matter of identifying duplicates and not updating them. In the case of trust attributes, however, there are a number of choices:
- don't update duplicate trust (shared database copy wins).
- overwrite trust from the legacy DB (legacy db copy wins).
- calculate the least common denominator trust between them (take the least trusted values). (turning off trust wins).
- calculate the most common demonimnator trust between the two (turning on trust wins).
From the user perspective, each of these choices means:
- after the update the application that just updated may trust certs that
it had previously marked untrusted, or certs that it has marked trusted are no longer trusted.
- after the update other applications that share the database may trust
certs they had previously marked untrusted, or certs that they had marked as trusted are no longer trusted.
- after the update all apps may find the certs that they marked trust are
no longer trusted.
- after the update all apps may find that they trust certs that have
previously been marked untrusted.
Option 3 is the most secure, Option 4 will break have less breakage. Trust merge conflicts that are real conflicts (application 'A' turned off trust and application 'B' turned on trust) are expected to be rare. The common case would be application 'A' turned on SSL trust and application 'B' turned on email trust. In this case Option 4 is clearly the correct choice.
From a programming point of view, NSS should pick a default and implement it. Ideally no user interaction will occur.
Finally password entries are merge issues. If the two databases have different passwords, the merged database will have to have a
Mozilla Applications
Mozilla applications are Mode 3A applications. In fact, for all intents and purposes, mozilla applications make the complete set of interesting 3A application. This section tries to map the issues above into actual code for Mozilla applications.
As 3A applications, Mozilla apps need to send NSS a unique identifier for the old cert and key database, as well as the old profile directory where the databases are stored. The profile 'salt' value would make a good unique identifier for mozilla products.
On startup, Mozilla apps should note when they are not in done state at initial nss startup (see flow chart for Mode 3A update above). If mozilla apps are not at 'done' state after startup, they should proceed to attempt to enter done state before PSM initialize completes.
Mozilla app will be in done state in the following cases (any of the below apply):
- The Mozilla app is starting as a fresh instance.
- The Mozilla app has already been updated.
- The shared database does not have a master password set and The legacy database for Mozilla app does a master password set.
These are the most common cases.
If the state is not done, then we know that this app has not already been updated, and either the shared database or the legacy database for the Mozilla app has a master password set.
UI question. At this point should we notify the user that we are updating the database to a shared database? In order to complete this we will need to do user interaction below.
If the legacy db for the mozilla app has a master password set, we prompt for it. This prompt must be clear we are asking for the master password for the running Mozilla app (Thunderbird, Firefox, Seamonkey, etc).
Exception case A
If we fail to get this password, we need to handle the exception A case. If the user has a master password set, but does not know what the master password is, then the following data is lost for sure:
The user's private keys. The user's secret keys. Any data encrypted to the private keys. Any data encrpted with the secret keys.
I believe we can identify if the private keys are associated with a certificate. If so, then we can tell the user what certificate would no longer work. Data encrypted with the private keys in Mozilla products are currently only email messages. Secret keys encrypt saved passwords. The Mozilla app knows which saved passwords are encrypted with that key.
If we hit Exception case A we can do one of the following:
1) attempt to just update the certs, trust, crl and s/mime records, skipping the all the keys. We would loose all the data described above. 2) decide not to update. In this case we would loose all the data in the paragraph above as well as all the certs, trust crl and s/mime records. 3) run with the legacy database and allow the user to update later. 4) run with the new shared database and allow the user to update later.
I would suggest we only offer the user the choice of 1 or 4. Note: if the user selects 1, the update could fail again in exception case B. From a UI perspective, we may want to handle exception case B as we handle case A so the user is only asked once about forcing an update while losing data.
Once we have a legacy db password, or if we determine we don't need the legacy
db password (either because there isn't one, or because we are willing to loose
the data that was protected by it). We need to acquire the shared db's
password so we can encrypt and mac the data properly. If the shared db doesn't
have a password we can proceed with the update without further prompting the
user. If the shared db has the same password as the legacy db, then we can
detect that and again proceed with the update without further prompting.
If both of these fail, we prompt for the password for the shared database. This prompt is trickier, because we need to ask the user for the password that he percieves to be the Master password for a different mozilla app. Note: at this point we are in a pretty uncommon corner case. Most users will not have different Master passwords for both Thunderbird and Firefox, for instance. However if we do arrive at this case, it is highly likely the user is not an experienced/informed user, so we need to treat this case carefully.
If we get the password, we complete the update as planned.
Exception case B
If we fail to get this password, we need to handle the exception B case. If the user has a master password set on his shared database, but does not know what that master password is, we now have the following choices:
1) eshew any private keys, secret keys and trust updates from the legacy database. 2) reset the password on the shared database (loosing all private and secret keys, possibly loosing some trust). 3) run with the legacy database and allow the user to update later. 4) run with the new shared database and allow the user to update later.
It seems pretty unlikely that the user truly does not know the shared database password, since he had to create or set it recently. However as the deployment time increases, this becomes more likely.
Again, I think giving the user a choice between options 1) and 4) are the best alternatives. If the user had already tripped over Exception case A, we can presume the user intends to make a similiar choice here. Case 2 can be handled later under the same way the user handles a forgotten master password today (only now resetting the master password affects all mozilla apps).
Profile issues
Mozilla apps can create more than one profile. Developers use this capability to test bugs that new users are likely to run into without loosing their own production evironment.
Shared databases, in general, means that some of the current semantics of user profiles will break. Creating a new profile will not create a new key/cert/master password profile. For developers (the primary users of profiles) it seems importan to preserve some of the existing semantics. I can see a couple of options.
- Allow profiles to be marked with 'private key/cert DB's. This will change
The Mozilla app from a Mode 3A app to a Mode 2A app. This will return developers to their previous semantic if they want, while allowing them to also test the interaction of different profiles and the same database. It would require UI changes to the profile manager, and it will require action on the part of the developer to get back to the old semantic.
- Treat only the default profile as Mode 3A and all other profiles as Mode 2A.
This will allow profile separation to operate as is today with no changes. It does mean, however, that only default profiles will share keys with appllication.
- Provide the checkbox in option 1, but make it default as in option 2.
I think option 3 probably provides the best solution for all worlds.
Database Merge
Layering
In order to keep clean separation between the data and database operations, we will continue to maintain an layer between the actual data handling and interpretation and the database itself. The database code will not need to understand:
- What objects are actually stored in it.
- The types of the attributes.
- The meaning of the stored data.
Softoken (not the database adapter layer) will manage canonicalizing any CK_ULONGs, encrypting or decrypting private data blobs, checking integrity and deciding what attributes an object should have and setting the appropriate defaults if necessary.
Since softoken deals with PKCS #11 templates internally, its interface to the database will be in terms of those templates.
The database layer must be multi-thread safe. If the underlying database is not thread safe, sdb_ layer must implement the appropriate locking.
s_open
The database API consists of an initialization call, which returns an SDB data structure (defined below).
CK_RV s_open(const char *directory, const char *certPrefix, const char *keyPrefix, int cert_version, int key_version, int flags, SDB **certdb, SDB **keydb, int *newInit)
The sdb_init function takes:
- directory full path to where the database lives.
- certPrefix a prefix string to add in front of the key and cert db (if keyPrefix is null), null means add no prefix.
- keyPrefix a prefix string to add in front of the key db. Null means use the same prefix as the cert db.
- cert_version current version is the current database version
- key_version is the current key database version
- flags are:
- FORCE
- READONLY
- READ/WRITE/CREATE
- certdb is the returned cert SDB structure
- keydb is the returned key SDB structure
- newInit returns 1 of s_open created new instances of cert and key (used for update).
The returned SDB structure has the following format:
typedef struct SDBStr SDB;
struct SDBStr { void *private; void *sdb_app_private; int sdb_type; int sdb_flags; int sdb_version; CK_RV (*sdb_FindObjectsInit)(SDB *sdb, const CK_ATTRIBUTE *template, int count, SDBFind **find); CK_RV (*sdb_FindObjects)(SDB *sdb, SDBFind *find, CK_OBJECT_HANDLE *ids, int arraySize, int *count); CK_RV (*sdb_FindObjectsFinal)(SDB *sdb, SDBFind *find); CK_RV (*sdb_GetAttributeValue)(SDB *sdb, CK_OBJECT_HANDLE object, CK_ATTRIBUTE *template, int count); CK_RV (*sdb_SetAttributeValue)(SDB *sdb, CK_OBJECT_HANDLE object, const CK_ATTRIBUTE *template, int count); CK_RV (*sdb_CreateObject)(SDB *sdb, CK_OBJECT_HANDLE *object, const CK_ATTRIBUTE *template, int count); CK_RV (*sdb_DestroyObject)(SDB *sdb, CK_OBJECT_HANDLE object); CK_RV (*sdb_GetPWEntry)(SDB *sdb, SDBPasswordEntry *entry); CK_RV (*sdb_PutPWEntry)(SDB *sdb, SDBPasswordEntry *entry); CK_RV (*sdb_Begin)(SDB *sdb); CK_RV (*sdb_Commit)(SDB *sdb); CK_RV (*sdb_Abort)(SDB *sdb); CK_RV (*sdb_Reset)(SDB *sdb); CK_RV (*sdb_Close)(SDB *sdb); };
where:
- private is a pointer to opaque private data specific to the Shared DB implementation.
- sdb_type is the type of database (key [aka private] or cert [aka public]).
- sdb_flags specifies how the database was opened (ReadOnly, Create, etc).
- sdb_version specifies the version of the underlying sdb structure. This allows us to handle future expansion of the sdb data structure safely.
- The rest are function pointers to database primitives described next.
sdb_FindObjectsInit
CK_RV (*sdb_FindObjectsInit)(SDB *sdb, const CK_ATTRIBUTE *template, int count, SDBFind **find);
This function is the equivalent of PKCS #11 C_FindObjectsInit(). It returns a SDBFind context with is opaque to the caller. The caller must call sdb_FindObjectsFinal with this context if sdb_FindobjectsInit succeeds.
sdb_FindObjects
CK_RV (*sdb_FindObjects)(SDB *sdb, SDBFind *find, CK_OBJECT_HANDLE *ids, int arraySize, int *count);
This function is the equivalent of PKCS #11 C_FindObjects(). It takes a SDBFind context returned by sdb_FindObjectsInit. This function has the same semantics as C_FindObjects with respect to handling how many objects are returned in a single call.
sdb_FindObjectsFinal
CK_RV (*sdb_FindObjectsFinal)(SDB *sdb, SDBFind *find);
This function is the equivalent of PKCS #11 C_FindObjectsFinal(). It frees any resources associated with SDBFIND.
sdb_GetAttributeValue
CK_RV (*sdb_GetAttributeValue)(SDB *sdb, CK_OBJECT_HANDLE object, CK_ATTRIBUTE *template, int count);
This function is the equivalent of PKCS #11 C_GetAttributeValue(). It has the same memory allocation and error code semantics of the PKCS #11 call. The attributes passed to sdb_GetAttributeValues are already transformed from their native representations in the following ways:
- CKU_LONG values are stored as 32-bit values in network byte order.
- Private attributes will be encrypted.
sdb_SetAttributeValue
CK_RV (*sdb_SetAttributeValue)(SDB *sdb, CK_OBJECT_HANDLE object, const CK_ATTRIBUTE *template, int count);
This function is the equivalent of PKCS #11 C_SetAttributeValue(). The attributes returned to sdb_SetAttributeValues are transformed from their native representations in the following ways:
- CKU_LONG values returned 32-bit values in network byte order.
- Private attributes returned encrypted.
sdb_CreateObject
CK_RV (*sdb_CreateObject)(SDB *sdb, CK_OBJECT_HANDLE *object, const CK_ATTRIBUTE *template, int count);
This function is the equivalent of PKCS #11 C_CreateObject(). The value of 'object' is chosen by the implementer of sdb_CreateObject. This value must be unique for this sdb instance. It should be no more than 30 bits long.
sdb_DestroyObject
CK_RV (*sdb_DestroyObject)(SDB *sdb, CK_OBJECT_HANDLE object);
This function is the equivalent of PKCS #11 C_Destroy object(). It removed the object from the database.
sdb_GetPWEntry
CK_RV (*sdb_GetPWEntry)(SDB *sdb, SDBPasswordEntry *entry);
Get the password entry. This only applies to the private database.
sdb_PutPWEntry
CK_RV (*sdb_PutPWEntry)(SDB *sdb, SDBPasswordEntry *entry);
Write the password entry. This only applies to the private database. Writing a password entry will overwrite the old entry.
sdb_Begin
CK_RV (*sdb_Begin)(SDB *sdb);
Begin a transaction. Any write to the database (sdb_CreateObject, sdb_DestroyObject, sdb_SetAttributeValue) must be accomplished while holding a transaction. Transactions are completed by calling sdb_Commit to commit the change, or sdb_Abort to discard the change. More than one write operation may be made while holding a transaction. Aborting the transaction will discard all writes made while in the transaction.
sdb_Commit
CK_RV (*sdb_Commit)(SDB *sdb);
Commit a transaction. Any write to the database (sdb_CreateObject, sdb_DestroyObject, sdb_SetAttributeValue) must be accomplished while holding a transaction. Transactions are completed by calling sdb_Commit to commit the change, or sdb_Abort to discard the change. More than one write operation may be made while holding a transaction.
sdb_Abort
CK_RV (*sdb_Abort)(SDB *sdb);
Abort a transaction. Any write to the database (sdb_CreateObject, sdb_DestroyObject, sdb_SetAttributeValue) must be accomplished while holding a transaction. Transactions are completed by calling sdb_Commit to commit the change, or sdb_Abort to discard the change. More than one write operation may be made while holding a transaction. Aborting the transaction will discard all writes made while in the transaction.
sdb_Close
CK_RV (*sdb_Close)(SDB *sdb);
Close the SDB and free up any resources associated with it.
sdb_Reset
CK_RV (*sdb_Reset)(SDB *sdb);
Reset zeros out the key database and resets the password.
legacy DB support
The old dbm code can be supported with the above SDB structure with the following exceptions:
- The old db code cannot be extensible (can't dynamically handle new types).
- A private interface may be needed to unwrap the private keys, or provide a handle to the password so the keys can be presented in the attribute format.
This code would live in its own shared library, called lgdbm (with the appropriate platform semantics, lgdbm.dll on windows, liblgdbm.so on unix, etc). Most of the low level cert, CRL, key handling, and translation to PKCS #11 objects and attributes that was part of softoken will moved to this legacy shared library. When access to old databased are needed, the lgdbm shared library will be loaded, and the following symbols will be dynamically found:
- legacy_Open - This has the same signature as s_open and returns SDB handles for the legacy database.
- legacy_ReadSecmodDB, legacy_ReleaseSecmodDBData, legacy_DeleteSecmodDB, legacy_AddSecmodDB - These functions provide access to the old secmod databases.
- legacy_Shutdown - This is called when NSS is through with all database support (that is when softoken shuts down).
- legacy_SetCryptFunctions - This is used to set some callbacks that the legacy db can call to decrypt and encrypt password protected records (pkcs8 formatted keys, etc.). This allows the legacy database to translate it's database records to the new format without getting direct access to the keys.
NSS will automaticall load the legacy database support under the following conditions:
- The application requests that the old databases be loaded (either implicitly or explicitly).
- The application request that new databases are loaded, but the new databases do not exist and the old databases do.