What is the most efficient way to read a large binary file python
问题 I have a large (21 GByte) file which I want to read into memory and then pass to a subroutine which processes the data transparently to me. I am on python 2.6.6 on Centos 6.5 so upgrading the operating system or python is not an option. Currently, I am using f = open(image_filename, "rb") image_file_contents=f.read() f.close() transparent_subroutine ( image_file_contents ) which is slow (~15 minutes). Before I start reading the file, I know how big the file is, because I call os.stat( image