The company I work for has an ATMOS cloud setup across multiple data centers.
ATMOS is an object-based cloud storage platform to store, archive and access unstructured content at scale.
We had an existing instance of Alfresco leveraging NetApp storage that I had to migrate to ATMOS.
In order to do this I needed to download and install 2 tools.
- ATMOS FOX - a firefox plugging that allows you to browse you ATMOS storage. https://addons.mozilla.org/en-US/firefox/addon/atmosfox/
- AtmosSync.jar - a tool to copy files existing on a filesystem to ATMOS. https://code.google.com/p/atmos-java/downloads/detail?name=AtmosSync.jar&can=2&q=label%3AOpSys-All
1. Stop your Alfresco and Solr servers
2. Create your S3 bucket with ATMOS FOX.
3. After you create the bucket you must add a non listable meta tag (bucket-mapping-type=one_to_one) to the bucket folder.
4. Copy your files in your contentstore and contentstore.deleted to Atmos using AtmosSync.jar.
5. Update your alfresco-global.properties with your S3 configuration.
### S3 Config ###
s3.accessKey=xxxxxxxxxxxx/xxxx
s3.secretKey=xxxxxxxxs3.bucketName=bucketNAME
#s3.bucketLocation=US
s3.flatRoot=falses3service.https-only=false
s3service.s3-endpoint=ATMOSHOST
s3service.s3-endpoint-http-port=8080#
s3service.disable-dns-buckets=false
dir.contentstore=contentstore
dir.contentstore.deleted=contentstore.deleted
#Maximum disk usage for the cache in MB
system.content.caching.maxUsageMB=51200
#Maximum size of files which can be stored in the cache in MB (zero implies no limit)
system.content.caching.maxFileSizeMB=0
6. Back up your DB if you haven’t already
7. Update all records in the ALF_CONTENT_URL table (store:// to s3://)
UPDATE alf_content_url SET content_url = replace(content_url, 'store:', 's3:’)
8. Startup Alfresco and Solr servers and you should be good to go.
Once you have verified that Alfresco is functioning properly you can repurpose the filesystem storage.
Hi Josh,
ReplyDeleteGood blog, nice and clear. We are trying to do the same and have migrated small amounts of content successfully. However we have more than 4TB of content to migrate and the performance we have seen with the tools we have tried so far (s3cmd, ATMOS S3) have been very slow and would take months to complete.
How did you find the performance with your tool, what throughput did you get?