Run any Python function on 1000 computers in 1 second.
Burla is the world's simplest cluster compute software. It's open-source, works with GPU's, custom containers, and up to 10,000 CPU's at once.

A fully fledged data-platform any team can learn in minutes:
Burla comes with a simple web-platform so your entire team can schedule jobs, create pipelines, scale machine learning systems, or other research efforts without weeks of onboarding or setup.

How it works:
Burla only has one function:
from burla import remote_parallel_map
my_inputs = [1, 2, 3]
def my_function(my_input):
print("I'm running on my own separate computer in the cloud!")
return my_input
return_values = remote_parallel_map(my_function, my_inputs)
With Burla, running code in the cloud feels the same as coding locally:
Anything you print appears in your local terminal.
Exceptions thrown in your code are thrown on your local machine.
Your local python packages are automatically synchronized with the cluster.
Responses are pretty quick, you can call a million simple functions in a couple seconds!
Attach big hardware to functions that need it:
Zero config files, just simple arguments like func_cpu
& func_ram
.
from xgboost import XGBClassifier
def train_model(hyper_parameters):
model = XGBClassifier(n_jobs=64, **hyper_parameters)
model.fit(training_inputs, training_targets)
remote_parallel_map(train_model, parameter_grid, func_cpu=64, func_ram=256)
Simple, flexible pipelines:
Nest remote_parallel_map
calls to build simple, massively parallel pipelines.
Use background=True
to schedule function calls that keep running after you close your laptop.
from burla import remote_parallel_map
def process_record(record):
# Pretend this does some math per-record!
return result
def process_file(file):
results = remote_parallel_map(process_record, split_into_records(file))
upload_results(results)
def process_files(files):
remote_parallel_map(process_file, files, func_ram=16)
remote_parallel_map(process_files, [files], background=True)
Run code in any Docker image, using the latest GPU's:
Public or private, just paste a URI to your image and hit start. Burla works with any linux based Docker image.

Try it now:
Enter your email below and we'll send you a free managed instance! Compute is on us. If you decide you like it, try self-hosting! Self-hosted Burla can be deployed with one command.
Questions? Schedule a call, or email [email protected]. We're always happy to talk.