Skip to content

HTTPS clone URL

Subversion checkout URL

You can clone with HTTPS or Subversion.

Download ZIP
Browse files

Add support for Gridfs

  • Loading branch information...
commit 5d11fb3b919347d92be870e4d71cd6e346f3d7ee 1 parent 6a10b35
@gmontard authored
View
13 README.md
@@ -1,3 +1,16 @@
+About this Fork
+=========
+
+This fork extends the storage fonctionnality to support Gridfs (Mongo).
+This version works with mongo v1.1.4.
+
+You should add this to your Gemfile :
+ gem "mongo", "1.1.4"
+ gem "bson", "1.1.4"
+ gem "bson_ext", "1.1.4", :require => "bson_ext/cbson"
+
+
+
Paperclip
=========
View
1  lib/paperclip/storage.rb
@@ -1,2 +1,3 @@
require "paperclip/storage/filesystem"
require "paperclip/storage/s3"
+require "paperclip/storage/gridfs"
View
191 lib/paperclip/storage/gridfs.rb
@@ -0,0 +1,191 @@
+module Paperclip
+ module Storage
+ # Amazon's S3 file hosting service is a scalable, easy place to store files for
+ # distribution. You can find out more about it at http://aws.amazon.com/s3
+ # There are a few S3-specific options for has_attached_file:
+ # * +s3_credentials+: Takes a path, a File, or a Hash. The path (or File) must point
+ # to a YAML file containing the +access_key_id+ and +secret_access_key+ that Amazon
+ # gives you. You can 'environment-space' this just like you do to your
+ # database.yml file, so different environments can use different accounts:
+ # development:
+ # access_key_id: 123...
+ # secret_access_key: 123...
+ # test:
+ # access_key_id: abc...
+ # secret_access_key: abc...
+ # production:
+ # access_key_id: 456...
+ # secret_access_key: 456...
+ # This is not required, however, and the file may simply look like this:
+ # access_key_id: 456...
+ # secret_access_key: 456...
+ # In which case, those access keys will be used in all environments. You can also
+ # put your bucket name in this file, instead of adding it to the code directly.
+ # This is useful when you want the same account but a different bucket for
+ # development versus production.
+ # * +s3_permissions+: This is a String that should be one of the "canned" access
+ # policies that S3 provides (more information can be found here:
+ # http://docs.amazonwebservices.com/AmazonS3/2006-03-01/RESTAccessPolicy.html#RESTCannedAccessPolicies)
+ # The default for Paperclip is :public_read.
+ # * +s3_protocol+: The protocol for the URLs generated to your S3 assets. Can be either
+ # 'http' or 'https'. Defaults to 'http' when your :s3_permissions are :public_read (the
+ # default), and 'https' when your :s3_permissions are anything else.
+ # * +s3_headers+: A hash of headers such as {'Expires' => 1.year.from_now.httpdate}
+ # * +bucket+: This is the name of the S3 bucket that will store your files. Remember
+ # that the bucket must be unique across all of Amazon S3. If the bucket does not exist
+ # Paperclip will attempt to create it. The bucket name will not be interpolated.
+ # You can define the bucket as a Proc if you want to determine it's name at runtime.
+ # Paperclip will call that Proc with attachment as the only argument.
+ # * +s3_host_alias+: The fully-qualified domain name (FQDN) that is the alias to the
+ # S3 domain of your bucket. Used with the :s3_alias_url url interpolation. See the
+ # link in the +url+ entry for more information about S3 domains and buckets.
+ # * +url+: There are three options for the S3 url. You can choose to have the bucket's name
+ # placed domain-style (bucket.s3.amazonaws.com) or path-style (s3.amazonaws.com/bucket).
+ # Lastly, you can specify a CNAME (which requires the CNAME to be specified as
+ # :s3_alias_url. You can read more about CNAMEs and S3 at
+ # http://docs.amazonwebservices.com/AmazonS3/latest/index.html?VirtualHosting.html
+ # Normally, this won't matter in the slightest and you can leave the default (which is
+ # path-style, or :s3_path_url). But in some cases paths don't work and you need to use
+ # the domain-style (:s3_domain_url). Anything else here will be treated like path-style.
+ # NOTE: If you use a CNAME for use with CloudFront, you can NOT specify https as your
+ # :s3_protocol; This is *not supported* by S3/CloudFront. Finally, when using the host
+ # alias, the :bucket parameter is ignored, as the hostname is used as the bucket name
+ # by S3.
+ # * +path+: This is the key under the bucket in which the file will be stored. The
+ # URL will be constructed from the bucket and the path. This is what you will want
+ # to interpolate. Keys should be unique, like filenames, and despite the fact that
+ # S3 (strictly speaking) does not support directories, you can still use a / to
+ # separate parts of your file name.
+ module S3
+ def self.extended base
+ begin
+ require 'aws/s3'
+ rescue LoadError => e
+ e.message << " (You may need to install the aws-s3 gem)"
+ raise e
+ end
+
+ base.instance_eval do
+ @s3_credentials = parse_credentials(@options[:s3_credentials])
+ @bucket = @options[:bucket] || @s3_credentials[:bucket]
+ @bucket = @bucket.call(self) if @bucket.is_a?(Proc)
+ @s3_options = @options[:s3_options] || {}
+ @s3_permissions = @options[:s3_permissions] || :public_read
+ @s3_protocol = @options[:s3_protocol] || (@s3_permissions == :public_read ? 'http' : 'https')
+ @s3_headers = @options[:s3_headers] || {}
+ @s3_host_alias = @options[:s3_host_alias]
+ unless @url.to_s.match(/^:s3.*url$/)
+ @path = @path.gsub(/:url/, @url)
+ @url = ":s3_path_url"
+ end
+ AWS::S3::Base.establish_connection!( @s3_options.merge(
+ :access_key_id => @s3_credentials[:access_key_id],
+ :secret_access_key => @s3_credentials[:secret_access_key]
+ ))
+ end
+ Paperclip.interpolates(:s3_alias_url) do |attachment, style|
+ "#{attachment.s3_protocol}://#{attachment.s3_host_alias}/#{attachment.path(style).gsub(%r{^/}, "")}"
+ end
+ Paperclip.interpolates(:s3_path_url) do |attachment, style|
+ "#{attachment.s3_protocol}://s3.amazonaws.com/#{attachment.bucket_name}/#{attachment.path(style).gsub(%r{^/}, "")}"
+ end
+ Paperclip.interpolates(:s3_domain_url) do |attachment, style|
+ "#{attachment.s3_protocol}://#{attachment.bucket_name}.s3.amazonaws.com/#{attachment.path(style).gsub(%r{^/}, "")}"
+ end
+ end
+
+ def expiring_url(time = 3600)
+ AWS::S3::S3Object.url_for(path, bucket_name, :expires_in => time )
+ end
+
+ def bucket_name
+ @bucket
+ end
+
+ def s3_host_alias
+ @s3_host_alias
+ end
+
+ def parse_credentials creds
+ creds = find_credentials(creds).stringify_keys
+ (creds[Rails.env] || creds).symbolize_keys
+ end
+
+ def exists?(style = default_style)
+ if original_filename
+ AWS::S3::S3Object.exists?(path(style), bucket_name)
+ else
+ false
+ end
+ end
+
+ def s3_protocol
+ @s3_protocol
+ end
+
+ # Returns representation of the data of the file assigned to the given
+ # style, in the format most representative of the current storage.
+ def to_file style = default_style
+ return @queued_for_write[style] if @queued_for_write[style]
+ filename = path(style)
+ extname = File.extname(filename)
+ basename = File.basename(filename, extname)
+ file = Tempfile.new([basename, extname])
+ file.binmode
+ file.write(AWS::S3::S3Object.value(path(style), bucket_name))
+ file.rewind
+ return file
+ end
+
+ def create_bucket
+ AWS::S3::Bucket.create(bucket_name)
+ end
+
+ def flush_writes #:nodoc:
+ @queued_for_write.each do |style, file|
+ begin
+ log("saving #{path(style)}")
+ AWS::S3::S3Object.store(path(style),
+ file,
+ bucket_name,
+ {:content_type => instance_read(:content_type),
+ :access => @s3_permissions,
+ }.merge(@s3_headers))
+ rescue AWS::S3::NoSuchBucket => e
+ create_bucket
+ retry
+ rescue AWS::S3::ResponseError => e
+ raise
+ end
+ end
+ @queued_for_write = {}
+ end
+
+ def flush_deletes #:nodoc:
+ @queued_for_delete.each do |path|
+ begin
+ log("deleting #{path}")
+ AWS::S3::S3Object.delete(path, bucket_name)
+ rescue AWS::S3::ResponseError
+ # Ignore this.
+ end
+ end
+ @queued_for_delete = []
+ end
+
+ def find_credentials creds
+ case creds
+ when File
+ YAML::load(ERB.new(File.read(creds.path)).result)
+ when String, Pathname
+ YAML::load(ERB.new(File.read(creds)).result)
+ when Hash
+ creds
+ else
+ raise ArgumentError, "Credentials are not a path, file, or hash."
+ end
+ end
+ private :find_credentials
+ end
+ end
+end
View
205 lib/paperclip/storage/s3.rb
@@ -1,162 +1,78 @@
module Paperclip
module Storage
- # Amazon's S3 file hosting service is a scalable, easy place to store files for
- # distribution. You can find out more about it at http://aws.amazon.com/s3
- # There are a few S3-specific options for has_attached_file:
- # * +s3_credentials+: Takes a path, a File, or a Hash. The path (or File) must point
- # to a YAML file containing the +access_key_id+ and +secret_access_key+ that Amazon
- # gives you. You can 'environment-space' this just like you do to your
- # database.yml file, so different environments can use different accounts:
- # development:
- # access_key_id: 123...
- # secret_access_key: 123...
- # test:
- # access_key_id: abc...
- # secret_access_key: abc...
- # production:
- # access_key_id: 456...
- # secret_access_key: 456...
- # This is not required, however, and the file may simply look like this:
- # access_key_id: 456...
- # secret_access_key: 456...
- # In which case, those access keys will be used in all environments. You can also
- # put your bucket name in this file, instead of adding it to the code directly.
- # This is useful when you want the same account but a different bucket for
- # development versus production.
- # * +s3_permissions+: This is a String that should be one of the "canned" access
- # policies that S3 provides (more information can be found here:
- # http://docs.amazonwebservices.com/AmazonS3/2006-03-01/RESTAccessPolicy.html#RESTCannedAccessPolicies)
- # The default for Paperclip is :public_read.
- # * +s3_protocol+: The protocol for the URLs generated to your S3 assets. Can be either
- # 'http' or 'https'. Defaults to 'http' when your :s3_permissions are :public_read (the
- # default), and 'https' when your :s3_permissions are anything else.
- # * +s3_headers+: A hash of headers such as {'Expires' => 1.year.from_now.httpdate}
- # * +bucket+: This is the name of the S3 bucket that will store your files. Remember
- # that the bucket must be unique across all of Amazon S3. If the bucket does not exist
- # Paperclip will attempt to create it. The bucket name will not be interpolated.
- # You can define the bucket as a Proc if you want to determine it's name at runtime.
- # Paperclip will call that Proc with attachment as the only argument.
- # * +s3_host_alias+: The fully-qualified domain name (FQDN) that is the alias to the
- # S3 domain of your bucket. Used with the :s3_alias_url url interpolation. See the
- # link in the +url+ entry for more information about S3 domains and buckets.
- # * +url+: There are three options for the S3 url. You can choose to have the bucket's name
- # placed domain-style (bucket.s3.amazonaws.com) or path-style (s3.amazonaws.com/bucket).
- # Lastly, you can specify a CNAME (which requires the CNAME to be specified as
- # :s3_alias_url. You can read more about CNAMEs and S3 at
- # http://docs.amazonwebservices.com/AmazonS3/latest/index.html?VirtualHosting.html
- # Normally, this won't matter in the slightest and you can leave the default (which is
- # path-style, or :s3_path_url). But in some cases paths don't work and you need to use
- # the domain-style (:s3_domain_url). Anything else here will be treated like path-style.
- # NOTE: If you use a CNAME for use with CloudFront, you can NOT specify https as your
- # :s3_protocol; This is *not supported* by S3/CloudFront. Finally, when using the host
- # alias, the :bucket parameter is ignored, as the hostname is used as the bucket name
- # by S3.
- # * +path+: This is the key under the bucket in which the file will be stored. The
- # URL will be constructed from the bucket and the path. This is what you will want
- # to interpolate. Keys should be unique, like filenames, and despite the fact that
- # S3 (strictly speaking) does not support directories, you can still use a / to
- # separate parts of your file name.
- module S3
+ # MongoDB's GridFS storage system (http://www.mongodb.org/display/DOCS/GridFS) uses
+ # a chunking strategy to store files in a mongodb database.
+ # Specific options for GridFS:
+ # * +gridfs_credentials+: Similar to +s3_credentials+, this can be a path, a File, or
+ # a Hash. Keys are as follows:
+ # * +database+: the name of the MongoDB database to connect to. This can also be
+ # a +Mongo::DB+ object, in which case that connection will be used, and other
+ # credentials will be ignored.
+ # * +host+: defaults to +localhost+
+ # * +username+ and +password+: optional authentication for the database server.
+ #
+ # Note that, because files stored using the +:gridfs+ storage module are stored
+ # within the database rather than the file system, you'll need to work out a method
+ # to extract the file data to serve it over HTTP. This is pretty trivial using
+ # Rails Metal.
+ module Gridfs
def self.extended base
begin
- require 'aws/s3'
+ require 'mongo'
rescue LoadError => e
- e.message << " (You may need to install the aws-s3 gem)"
+ e.message << " (You may need to install the mongo gem)"
raise e
end
-
+
base.instance_eval do
- @s3_credentials = parse_credentials(@options[:s3_credentials])
- @bucket = @options[:bucket] || @s3_credentials[:bucket]
- @bucket = @bucket.call(self) if @bucket.is_a?(Proc)
- @s3_options = @options[:s3_options] || {}
- @s3_permissions = @options[:s3_permissions] || :public_read
- @s3_protocol = @options[:s3_protocol] || (@s3_permissions == :public_read ? 'http' : 'https')
- @s3_headers = @options[:s3_headers] || {}
- @s3_host_alias = @options[:s3_host_alias]
- unless @url.to_s.match(/^:s3.*url$/)
- @path = @path.gsub(/:url/, @url)
- @url = ":s3_path_url"
- end
- AWS::S3::Base.establish_connection!( @s3_options.merge(
- :access_key_id => @s3_credentials[:access_key_id],
- :secret_access_key => @s3_credentials[:secret_access_key]
- ))
- end
- Paperclip.interpolates(:s3_alias_url) do |attachment, style|
- "#{attachment.s3_protocol}://#{attachment.s3_host_alias}/#{attachment.path(style).gsub(%r{^/}, "")}"
+ @gridfs_options = parse_credentials(@options[:gridfs])
+ @gridfs_db = Paperclip::Storage::Gridfs.gridfs_connections(@gridfs_options)
end
- Paperclip.interpolates(:s3_path_url) do |attachment, style|
- "#{attachment.s3_protocol}://s3.amazonaws.com/#{attachment.bucket_name}/#{attachment.path(style).gsub(%r{^/}, "")}"
- end
- Paperclip.interpolates(:s3_domain_url) do |attachment, style|
- "#{attachment.s3_protocol}://#{attachment.bucket_name}.s3.amazonaws.com/#{attachment.path(style).gsub(%r{^/}, "")}"
- end
- end
-
- def expiring_url(time = 3600)
- AWS::S3::S3Object.url_for(path, bucket_name, :expires_in => time )
end
-
- def bucket_name
- @bucket
+
+ def self.gridfs_connections creds
+ @connections ||= {}
+ @connections[creds] ||= get_database_connection(creds)
end
-
- def s3_host_alias
- @s3_host_alias
+
+ def connection
+ @gridfs_db
end
-
+
def parse_credentials creds
creds = find_credentials(creds).stringify_keys
- (creds[Rails.env] || creds).symbolize_keys
+ (creds[RAILS_ENV] || creds).symbolize_keys
end
-
+
def exists?(style = default_style)
if original_filename
- AWS::S3::S3Object.exists?(path(style), bucket_name)
+ @gridfs = Mongo::GridFileSystem.new(connection)
+ val = @gridfs.open(path(style), "r") rescue nil
+ !val.nil?
else
false
end
end
-
- def s3_protocol
- @s3_protocol
- end
-
+
# Returns representation of the data of the file assigned to the given
# style, in the format most representative of the current storage.
def to_file style = default_style
- return @queued_for_write[style] if @queued_for_write[style]
- filename = path(style)
- extname = File.extname(filename)
- basename = File.basename(filename, extname)
- file = Tempfile.new([basename, extname])
- file.binmode
- file.write(AWS::S3::S3Object.value(path(style), bucket_name))
- file.rewind
- return file
- end
-
- def create_bucket
- AWS::S3::Bucket.create(bucket_name)
+ @queued_for_write[style] || (Mongo::GridFileSystem.open(connection, path(style), 'rb') if exists?(style))
end
def flush_writes #:nodoc:
- @queued_for_write.each do |style, file|
- begin
- log("saving #{path(style)}")
- AWS::S3::S3Object.store(path(style),
- file,
- bucket_name,
- {:content_type => instance_read(:content_type),
- :access => @s3_permissions,
- }.merge(@s3_headers))
- rescue AWS::S3::NoSuchBucket => e
- create_bucket
- retry
- rescue AWS::S3::ResponseError => e
- raise
- end
+ @queued_for_write.each do |style, file|
+ log("saving #{path(style)}")
+ @gridfs = Mongo::GridFileSystem.new(connection)
+ @gridfs.open(path(style), 'w', {
+ :content_type => content_type,
+ :metadata => { 'instance_id' => instance.id },
+ :chunk_size => 4.kilobytes
+ }) { |f|
+ f.write file.read
+ }
+ file.close
+ File.unlink(file.path)
end
@queued_for_write = {}
end
@@ -165,19 +81,35 @@ def flush_deletes #:nodoc:
@queued_for_delete.each do |path|
begin
log("deleting #{path}")
- AWS::S3::S3Object.delete(path, bucket_name)
- rescue AWS::S3::ResponseError
- # Ignore this.
+ @gridfs = Mongo::GridFileSystem.new(connection)
+ val = @gridfs.open(path, "r") rescue nil
+ if !val.nil?
+ @gridfs.delete(path)
+ end
+ rescue Errno::ENOENT => e
+ # ignore file-not-found, let everything else pass
end
end
@queued_for_delete = []
end
+
+ def self.get_database_connection creds
+ case creds[:database]
+ when Mongo::DB then creds[:database]
+ else
+ returning Mongo::Connection.new(creds[:host] || "localhost", creds[:port] || Mongo::Connection::DEFAULT_PORT).db(creds[:database]) do |db|
+ if creds[:username] && creds[:password]
+ auth = db.authenticate creds[:username], creds[:password]
+ end
+ end
+ end
+ end
def find_credentials creds
case creds
when File
YAML::load(ERB.new(File.read(creds.path)).result)
- when String, Pathname
+ when String
YAML::load(ERB.new(File.read(creds)).result)
when Hash
creds
@@ -186,7 +118,6 @@ def find_credentials creds
end
end
private :find_credentials
-
end
end
end
Please sign in to comment.
Something went wrong with that request. Please try again.