run_task comes as a part of dsci SDK for Python . What do you mean by build time dependencies? You are free put anything into Python / Bash tasks that are called by run_task function in job file … if you point your repo on demo server I can help you with that …
melezhik
Update for new users: Just create some repo and add http://127.0.0.1:4000/forgejo_hook to a repo web hook, then create some dsci pipeline and it will gets triggered
Actually just enabled self registration , no need to ask me, but still will be happy to see anyone in discord )
It uses yaml only for configuration part, but pipeline itself is far more then that . Not sure what do you mean by “middle ground”, could you please elaborate? Thanks
Just added the feature of running jobs on localhost for debugging:
cd .dsci/job_one; docker run -it -v $PWD:/opt/job --entrypoint /bin/bash dsci -c "cd /opt/job/; s6 --task-run ."
Feedback are welcome , the project is in very early stage …
TLTR;
Sparky is a distributed jobs framework that allows orchestration of remote tasks on cluster of nodes. It’s simple to set up and easy to use. This post is a brief overview of Sparky architecture design.
Sparky targeted audience is:
cloud providers to manage underlying multiple hosts infrastructure
data scientists to process data in distributed manner (aka data pipelines)
software engineers and devops doing any tasks with distributed nature in mind
You may try out https://github.com/melezhik/sparky which is a local / remote task runner with nice front end and scripts could be written on many languages
Yep. Fancy devs watching me coding some Rakulang in nano 😂
By default this is alline:latest docker container , however one can use custom docker images ( to be documented, but let me know if you are interested ), on the demo server , Ubuntu image is used