Create some test files and check them in a large loop to see the average time. Make sure the test files match your scripts.
I used this code:
import glob import time times1 = [] for i in range(0,1000): names = {} t0 = time.clock() with open("lines.txt") as f: names["lines.txt"] = sum(1 for line in f if line.strip()) print names times1.append(time.clock()-t0) times2 = [] for i in range(0,1000): names = {} t0 = time.clock() data = open("lines.txt", 'r').read() print("lines.txt",len(data.splitlines()), len(data.split()), len(data)) times2.append(time.clock()-t0) print sum(times1)/len(times1) print sum(times2)/len(times2)
and left with average timings: 0.0104755582104 and 0.0180650466201 seconds
It was a text file with 23,000 lines. For instance:
print("lines.txt",len(data.splitlines()), len(data.split()), len(data))
outputs: ('lines.txt', 23056, 161392, 1095160)
Check this out on your actual fileset to get more accurate sync data.
Pedro source share