Mongoose storage driver extention for testing of S3 type storages. The repo contains only the extension source code, the source code of the mongoose core and the full mongoose documentation is contained in the mongoose-base
repository.
- Features
- Deployment
2.1. Jar
2.2. Docker - Configuration Reference
3.1. S3 Specific Options
3.2. Other Options - Usage
4.1. Main functionality
4.1. HTTP functionality
4.2. Object Tagging
4.3. Versioning>/br> - Minio S3 server
- API version: 2006-03-01
- Authentification:
- SSL/TLS
- Item types:
data
(--> "object")path
(--> "bucket")
- Automatic destination path creation on demand
- Path listing input (with XML response payload)
- Data item operation types:
create
read
- full
- random byte ranges
- fixed byte ranges
- content verification
- object tagging
update
- full (overwrite)
- random byte ranges
- fixed byte ranges (with append mode)
- object tagging
delete
- full
- object tagging
noop
- Path item operation types:
create
read
delete
noop
Java 11+ is required to build/run.
-
Get the latest
mongoose-base
jar from the maven repo and put it to your working directory. Note the particular version, which is referred as BASE_VERSION below. -
Get the latest
mongoose-storage-driver-coop
jar from the maven repo and put it to the~/.mongoose/<BASE_VERSION>/ext
directory. -
Get the latest
mongoose-storage-driver-netty
jar from the maven repo and put it to the~/.mongoose/<BASE_VERSION>/ext
directory. -
Get the latest
mongoose-storage-driver-http
jar from the maven repo and put it to the~/.mongoose/<BASE_VERSION>/ext
directory. -
Get the latest
mongoose-storage-driver-s3
jar from the maven repo and put it to the~/.mongoose/<BASE_VERSION>/ext
directory.
java -jar mongoose-base-<BASE_VERSION>.jar \
--storage-driver-type=s3 \
[<MONGOOSE CLI ARGS>]
NOTE: The base image doesn't contain any additonal load step types neither additional storage drivers.
Example:
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-net-node-addrs=<NODE_IP_ADDRS> \
[<MONGOOSE CLI ARGS>]
Example:
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--run-node
NOTE: Mongoose uses
1099
port for RMI between mongoose nodes and9999
for REST API. If you run several mongoose nodes on the same host (in different docker containers, for example) or if the ports are used by another service, then ports can be redefined:docker run \ --network host \ emcmongoose/mongoose-storage-driver-s3 \ --run-node \ --load-step-node-port=<RMI PORT> \ --run-port=<REST PORT>
Example:
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--load-step-node-addrs=<ADDR1,ADDR2,...> \
--storage-net-node-addrs=<NODE_IP_ADDRS> \
[<MONGOOSE CLI ARGS>]
Name | Type | Default Value | Description |
---|---|---|---|
storage-auth-version | Int | 2 | Specifies which auth version to use. Valid values: 2, 4. |
storage-object-fsAccess | Flag | false | Specifies whether filesystem access is enabled or not |
storage-object-tagging-enabled | Flag | false | Work (PUT/GET/DELETE) with object tagging or not (default) |
storage-object-tagging-tags | Map | {} | Map of name-value tags, effective only for the UPDATE operation when tagging is enabled |
storage-object-versioning | Flag | false | Specifies whether the versioning storage feature is used or not |
- A bucket may be specified with either
item-input-path
oritem-output-path
configuration option - Multipart upload should be enabled using the
item-data-ranges-threshold
configuration option - The default storage port is set to 9020 for the docker image
Examples of mongoose core usage
NOTE: Mongoose S3 SD depends on Mongoose HTTP SD, and the S3 bundle includes all the features of HTTP SD, so all http-specific parameters can be also used with this S3 driver.
Examples of HTTP headers usage
https://docs.aws.amazon.com/AmazonS3/latest/dev/object-tagging.html
https://docs.aws.amazon.com/AmazonS3/latest/API/API_PutObjectTagging.html
Put (create or replace) the tags on the existing objects. The update
load operation should be used for this.
The objects should be specified by an
item input
(the bucket listing or the items input CSV file).
Scenario example:
var updateTaggingConfig = {
"storage" : {
"object" : {
"tagging" : {
"enabled" : true,
"tags" : {
"tag0" : "value_0",
"tag1" : "value_1",
// ...
"tagN" : "value_N"
}
}
}
}
};
UpdateLoad
.config(updateTaggingConfig)
.run();
Command line example:
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-auth-uid=user1 \
--storage-auth-secret=**************************************** \
--item-input-file=objects_to_update_tagging.csv \
--item-output-path=/bucket1 \
--storage-net-transport=nio \
--run-scenario=tagging.js
Note:
It's not possible to use the command line to specify the tag set, a user should use the scenario file for this
Both tag names and values support the expression language:
Example:
var updateTaggingConfig = {
"storage" : {
"object" : {
"tagging" : {
"enabled" : true,
"tags" : {
"foo${rnd.nextInt()}" : "bar${time:millisSinceEpoch()}",
"key1" : "${date:formatNowIso8601()}",
"${e}" : "${pi}"
}
}
}
}
};
UpdateLoad
.config(updateTaggingConfig)
.run();
https://docs.aws.amazon.com/AmazonS3/latest/API/API_GetObjectTagging.html
Example:
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-net-node-addrs=<NODE_IP_ADDRS> \
--read \
--item-input-path=/bucket1 \
--storage-object-tagging-enabled \
[<MONGOOSE CLI ARGS>]
https://docs.aws.amazon.com/AmazonS3/latest/API/API_DeleteObjectTagging.html
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-net-node-addrs=<NODE_IP_ADDRS> \
--delete \
--item-input-file=objects_to_delete_tagging.csv \
--storage-object-tagging-enabled \
[<MONGOOSE CLI ARGS>]
Create request is the only versioning operation that doesn't require version-id by default. It just creates a new
version for the same object. We can only retrieve the version-ids by specifying item-input-file
. To create such file
make sure to enable --storage-object-versioning
flag and specify item-output-file
path.
There are two approaches to do load testing with versioning. But first stage is common:
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-auth-uid=user1 \
--storage-auth-secret=**************************************** \
--storage-net-node-addrs=<NODE_IP_ADDRS> \
--item-output-file=itemsInitialList.csv \
--storage-object-versioning=true \
--item-output-path=/bucket \
--load-op-limit-count=<N>
We create an itemsInitialList.csv
that has objects that we are going to version. Next step is different for the
two approaches.
First approach is to use --load-op-recycle
mode. We pass the list of objects to version and specify the
limit-count=<N*M>
where N
- is the length of the intial list and M
is the amount of versions per object. Be aware
that recycle-mode
doesn't guarantee the exact amount of versions per object. But the average amount will be M
.
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-auth-uid=user1 \
--storage-auth-secret=**************************************** \
--storage-net-node-addrs=<NODE_IP_ADDRS> \
--item-input-file=itemsInitialList.csv \
--storage-object-versioning=true \
--load-op-limit-count=<N*M>
There is a constraint. You need to have a large enough input-file so that Mongoose always has work to do in order to get max throughput. Because it can't send a new request on the same item until it's acked. And if we consider let's say 100ms latency for a small object then Mongoose should have enough objects to process during those 100ms to not stay idle.
To determine that amount check your latency and throughput when doing regular s3 PUTs. If you do 50000 op/s and your latency is 100ms then the initial list must be 5000 objects.
As a side-note: if you want to get a list of items for this test (e.g. to pass it to read test) make sure to
enable load-op-output-duplicates
flag as by default mongoose doesn't print the duplicates created by recycle mode.
Another approach requires using command line tools after the common step but guarantees the exact amount of versions per object. Instead of recycling object we can provide Mongoose a list of objects which would already have copies of the same object. This can be achieved for example via:
for i in {1..1000}; do cat itemsInitialList.csv; done > itemsWithVersionsList.csv
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-auth-uid=user1 \
--storage-auth-secret=**************************************** \
--storage-net-node-addrs=<NODE_IP_ADDRS> \
--item-input-file=itemsWithVersionsList.csv \
--storage-object-versioning=true
Unlike PUT operations, GETs are simple. You need to have an input-file
with versions like this generated by
PUT load:
/bucket/97bdgavrlkp0~161458,10cf8e8ba060d304,100,0/0
/bucket/gx8zqoy6fvtd~161494,1ee9b7ffddbddcd1,100,0/0
/bucket/vs71f8k3cnx9~161504,3a0e47edc025f71d,100,0/0
/bucket/hg9ai03dthjm~161516,1fe09f7b49e09712,100,0/0
/bucket/m3stkc24nmdp~161517,2860e144acfb5d0d,100,0/0
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-auth-uid=user1 \
--load-op-type=read \
--storage-auth-secret=**************************************** \
--storage-net-node-addrs=<NODE_IP_ADDRS> \
--item-input-file=itemsWithVersionsList.csv \
--storage-object-versioning=true
DELETEs are also simple. An input-file
is again required.
docker run \
--network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-auth-uid=user1 \
--load-op-type=delete \
--storage-auth-secret=**************************************** \
--storage-net-node-addrs=<NODE_IP_ADDRS> \
--item-input-file=itemsWithVersionsList.csv \
--storage-object-versioning=true
For tests, a minio/minio
S3 server is used.
It can be deployed to test the mongoose commands and S3-specific scenarios if there is no access to real S3 storage.
Example:
docker run -d --name s3_server \
-p 9000:9000 \
--env MINIO_ACCESS_KEY=user1 \
--env MINIO_SECRET_KEY=secretKey1 \
minio/minio:latest \
server /data
Mongoose run:
docker run --network host \
emcmongoose/mongoose-storage-driver-s3 \
--storage-net-node-port=9000 \
--storage-auth-uid=user1 \
--storage-auth-secret=secretKey1 \
--storage-net-node-addrs=localhost