Carbonio Storages and S3 buckets
The majority of your data may be stored in secure and long-lasting cloud storage thanks to
Carbonio Storages and S3 buckets,
Services compatible with S3
- (Standard Local Volume) FileBlob
- S3 Amazon
- S3 Scality
- Any unsupported S3-compliant solution that is custom
The “Incoming” Directory and Primary Volumes
; the following commands allow you to view or change the current setting:
Setup of a bucket
- an S3 container. To utilise the bucket, you must be aware of its name and location.
- Access Key and Secret of a user
- a rule that gives the user complete access to your bucket.
Management of buckets
Bucket Paths and Naming
- There can be as many volumes under the same destination route as you like because the Bucket Name and Destination route are not linked to the volume itself.
- On the other hand, the Volume Prefix is unique to each volume and serves as a rapid means of differentiating and identifying various volumes inside the bucket.
Amazon S3 Bucket of Tips
is required in order to receive an Access Key and the associated Secret. For simpler management, we advise you to create a dedicated user in Amazon’s IAM Service.
section to: if you simply want to grant the bare minimum of permissions.
According to Amazon’s standard naming convention, the bucket’s ARN is written as arn:partition:service:region:account-id:resource. Please refer to Amazon’s documentation for further details on this subject.
Bucket Paths and Naming
Infrequent Access Storage Class
value to this storage class. Carbonio Storages is compatible with the
Intelligent Tiering Storage Class
The definition of item deduplication
Item Deduplication Carbonio
performs item deduplication when a new item is stored in the Current Primary Volume.
of a newly produced item is compared to a list of previously cached objects. Instead of creating a brand-new BLOB for the message if there is a match, a hard link to the BLOB of the cached message is generated.
Deduplication of items with Carbonio storage systems
action of the Carbonio Storages parses a target volume to identify and remove any duplicated items.
procedure following a migration or a significant data import is also strongly advised.
Volume Deduplication being used
operation while it is executing by using the command. The doDeduplicate operation is a suitable target for the
command. Typical Output is:
- Current Pass (Prefix for Digest): Based on the first character of their digests (name), the BLOBS will be analysed by the
command in groups.
- Number of mailboxes that were checked during the current pass.
- Deduplicated/duplicated Blobs: Total number of duplicated items on the volume / Number of BLOBS deduplicated by the current operation
- Number of already deduplicated blobs (duplicated blobs that have already undergone a prior run) on the volume
- Blobs that have not been examined, typically as a result of a read error or a missing file, are known as skipped blobs.
- Invalid Digests: BLOBs having incorrect digests (names that differ from the file’s real digest).
- Amount of disc space saved overall with the doDeduplicate process.
- On the last mailbox, the procedure is now on the second-to-last pass.
- 137089 duplicated BLOBs have been discovered, of which 71178 have already undergone deduplication.
- 64868 BLOBs were deduplicated in the current operation, saving a total of 21.88GB in disc space.