Module: TensorFlow::Keras::Utils

Defined in:
lib/tensorflow/keras/utils.rb

Class Method Summary collapse

Class Method Details

.get_file(fname, origin, file_hash: nil, cache_subdir: "datasets") ⇒ Object



5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
# File 'lib/tensorflow/keras/utils.rb', line 5

def get_file(fname, origin, file_hash: nil, cache_subdir: "datasets")
  # destination
  # TODO handle this better
  raise "No HOME" unless ENV["HOME"]
  dest = "#{ENV["HOME"]}/.keras/#{cache_subdir}/#{fname}"
  FileUtils.mkdir_p(File.dirname(dest))

  return dest if File.exist?(dest)

  temp_dir ||= File.dirname(Tempfile.new("tensorflow"))
  temp_path = "#{temp_dir}/#{Time.now.to_f}" # TODO better name

  digest = file_hash&.size == 32 ? Digest::MD5.new : Digest::SHA2.new

  uri = URI(origin)

  # Net::HTTP automatically adds Accept-Encoding for compression
  # of response bodies and automatically decompresses gzip
  # and deflateresponses unless a Range header was sent.
  # https://ruby-doc.org/stdlib-2.6.4/libdoc/net/http/rdoc/Net/HTTP.html
  Net::HTTP.start(uri.host, uri.port, use_ssl: true) do |http|
    request = Net::HTTP::Get.new(uri)

    puts "Downloading data from #{origin}"
    i = 0
    File.open(temp_path, "wb") do |f|
      http.request(request) do |response|
        response.read_body do |chunk|
          f.write(chunk)
          digest.update(chunk)

          # print progress
          putc "." if i % 50 == 0
          i += 1
        end
      end
      puts # newline
    end
  end

  if file_hash && digest.hexdigest != file_hash
    raise Error, "Bad hash"
  end

  FileUtils.mv(temp_path, dest)

  dest
end