I have a large amount of binary data that needs to be pushed across the network. It appears from profiling that the dominant time is being taken up by packing the data. (50%) Here is a CME that shows the problem.
from numpy import random from struct import pack import time lenstim = 10084200 sigma = 0.1 stim = random.normal(0., sigma, lenstim) # 10084200 gaussian random numbers (doubles) fmt = '!h'+str(stim.size)+'d' # makes fmt = '!h10084200d' cmd = 4 startTime = time.time() packdat = pack( fmt, cmd, *stim ) elapsed = time.time() - startTime print "Time to pack the command and data %.6f seconds " % elapsed Is there a faster method to do this? Is it possible to use array? Any suggestions would be appreciated. -- http://mail.python.org/mailman/listinfo/python-list