a little framework to simule multiprocessing over a lot of computers
Project description
import random
import org.py4grid.GP as gp
def testando(lista):
for x in range(2 * 1):
lista.append(random.random())
return lista
if __name__ == '__main__':
from multiprocessing.dummy import Pool as pool
remote = gp.RemoteProcess(pool, file=__file__)
ret = remote.processwork(__file__, testando, [[], [], [], []],
relative_path={'Darwin': '/Dropbox/BIBLIOTECA_PYTHON', 'Linux': '/Dropbox/BIBLIOTECA_PYTHON'},
hosts=[('localhost', 4680)])
for item in ret:
for sub in item:
print(len(sub), sub)
before use this framework,
starts PY4GRIDSERVER generated in PythonXX\Script directory for begin a server
example:
C:\Python33\Script\PY4GRIDSERVER <- starts the server on port 4680
import org.py4grid.GP as gp
def testando(lista):
for x in range(2 * 1):
lista.append(random.random())
return lista
if __name__ == '__main__':
from multiprocessing.dummy import Pool as pool
remote = gp.RemoteProcess(pool, file=__file__)
ret = remote.processwork(__file__, testando, [[], [], [], []],
relative_path={'Darwin': '/Dropbox/BIBLIOTECA_PYTHON', 'Linux': '/Dropbox/BIBLIOTECA_PYTHON'},
hosts=[('localhost', 4680)])
for item in ret:
for sub in item:
print(len(sub), sub)
before use this framework,
starts PY4GRIDSERVER generated in PythonXX\Script directory for begin a server
example:
C:\Python33\Script\PY4GRIDSERVER <- starts the server on port 4680