I am doing a project for a company, they want a causal inference model to be made that reads from Google Sheets via BigQuery, performs a computation (using either Python, R, TensorFlow Jupyter notebook – I'm currently looking at the various options), then outputs the results to Data Studio.
- An instance of R or Python will need to be loaded up every time, should I use R or Python? I think Python has more packages for ML but need to do more research.
- Those instances need to be somewhere, so if its R I can create an instance of RStudio on a VM, or create an instance of Jupyter notebook on a VM, if its Python, I can use Jupyter only (I think) on a VM. I created an instance of RStudio but it seemed very expensive. some of the considerations are that they will be recurring use cases with a company budget available, and the easier for non technical people to understand the better.
- They want there to be an easy user interface so their less techy guys can upload data tables into the model, tweak some parameters, and get a result. I am thinking Jupyter notebook lends itself to this, but I also get the idea that Jupyter notebook is for experimenting and not necessarily designed to house the end product. Can anyone give an insight into this?
Could people suggest what route to go down and some of the pros and cons of each?
submitted by /u/Flewizzle
[link] [comments]
from Software Development – methodologies, techniques, and tools. Covering Agile, RUP, Waterfall + more! https://ift.tt/3gqcpKz