module PhusionPassenger::Utils::Download

def self.included(klass)

def self.included(klass)
  # When included into another class, make sure that Utils
  # methods are made private.
  public_instance_methods(false).each do |method_name|
    klass.send(:private, method_name)
  end
end

def basename_from_url(url)

def basename_from_url(url)
  return url.sub(/.*\//, '')
end

def download(url, output, options = {})

Set to nil to disable this timeout. Default: nil.
operation, including connection time. Only has effect on curl.
total_timeout: The maximum amount of time spent on the whole download
timeout. Default: 5.
to the default wget value, 900. Set to nil to disable this
idle_timeout: The maximum idle read time. Set to nil to set this timeout
disable this timeout. Default: 4.
and establishing the TCP connection. Set to nil to
connect_timeout: The maximum amount of time to spend on DNS lookup
Default: false.
use_cache: Whether to copy the file from the download cache, if available.
The default is to use the download tool's down CA database.
cacert: a CA certificate file to use for verifying SSL websites.
logger: the logger to use. If not given, this function will log to STDERR.
show_progress: whether to show download progress. Default: false.

Options:

Returns whether the download succeeded.
Downloads a file from the given URL and saves it to the given filename.
def download(url, output, options = {})
  options = {
    :connect_timeout => 4,
    :idle_timeout    => 5
  }.merge(options)
  logger = options[:logger] || Logger.new(STDERR)
  if options[:use_cache] && cache_dir = PhusionPassenger.download_cache_dir
    basename = basename_from_url(url)
    if File.exist?("#{cache_dir}/#{basename}")
      logger.info "Copying #{basename} from #{cache_dir}..."
      FileUtils.cp("#{cache_dir}/#{basename}", output)
      return true
    end
  end
  if PlatformInfo.find_command("curl")
    return download_with_curl(logger, url, output, options)
  elsif PlatformInfo.find_command("wget")
    return download_with_wget(logger, url, output, options)
  else
    logger.error "Could not download #{url}: no download tool found (curl or wget required)"
    return false
  end
end

def download_with_curl(logger, url, output, options)

def download_with_curl(logger, url, output, options)
  command = ["curl", "-f", "-L", "-o", output]
  if options[:show_progress]
    command << "-#"
  else
    command << "-s"
    command << "-S"
  end
  if options[:cacert]
    command << "--cacert"
    command << options[:cacert]
  end
  if options[:connect_timeout]
    command << "--connect-timeout"
    command << options[:connect_timeout].to_s
  end
  if options[:idle_timeout]
    command << "--speed-time"
    command << options[:idle_timeout].to_s
    command << "--speed-limit"
    command << "1"
  end
  if options[:total_timeout]
    command << "--max-time"
    command << options[:total_timeout].to_s
  end
  command << url
  command_str = Shellwords.join(command)
  logger.info("Invoking: #{command_str}")
  if options[:show_progress]
    # If curl errors out we don't want it to display 'curl: ' prefixes,
    # so we parse its output.
    begin
      io = IO.popen("#{command_str} 2>&1", "r")
    rescue SystemCallError => e
      logger.error("Could not invoke curl: #{e}")
      return false
    end
    begin
      non_empty_line_encountered = false
      while !io.eof?
        # We split on "\r" because progress bar lines do not contain "\n".
        data = io.gets("\r")
        data = remove_curl_output_prefix(data)
        # If an error occurs then the first few lines may be empty.
        # Skip those.
        if !non_empty_line_encountered && data =~ /\A\n+/
          data.gsub!(/\A\n+/, '')
        end
        non_empty_line_encountered = true
        STDERR.write(data)
        STDERR.flush
      end
    ensure
      io.close
    end
    result = $?.exitstatus == 0
  else
    begin
      output = `#{command_str} 2>&1`
    rescue SystemCallError => e
      logger.error("Could not invoke curl: #{e}")
      return false
    end
    result = $?.exitstatus == 0
    if !result
      output = remove_curl_output_prefix(output)
      output.chomp!
      logger.error("Could not download #{url}: #{output}")
    end
  end
  return result
end

def download_with_wget(logger, url, output, options)

def download_with_wget(logger, url, output, options)
  command = ["wget", "--tries=1", "-O", output]
  if !options[:show_progress]
    command << "-nv"
  end
  if options[:cacert]
    command << "--ca-certificate=#{options[:cacert]}"
  end
  if options[:connect_timeout]
    command << "--dns-timeout=#{options[:connect_timeout]}"
    command << "--connect-timeout=#{options[:connect_timeout]}"
  end
  if options[:idle_timeout]
    command << "--timeout=#{options[:idle_timeout]}"
  end
  command << url
  command_str = Shellwords.join(command)
  logger.info("Invoking: #{command_str}")
  if options[:show_progress]
    begin
      result = system(*command)
    rescue SystemCallError => e
      logger.error("Could not invoke wget: #{e}")
      return false
    end
    if !result
      logger.error("Could not download #{url}: #{output}")
    end
  else
    begin
      output = `#{command_str} 2>&1`
    rescue SystemCallError => e
      logger.error("Could not invoke wget: #{e}")
      return false
    end
    result = $?.exitstatus == 0
    if !result
      # Error output may begin with "<URL>:\n" which is redundant.
      output.gsub!(/\A#{Regexp.escape url}:\n/, '')
      output.chomp!
      logger.error("Could not download #{url}: #{output}")
    end
  end
  return result
end

def remove_curl_output_prefix(line)

def remove_curl_output_prefix(line)
  return line.gsub(/^curl: (\([0-9]+\) )?/, '')
end