This repository has been archived by the owner on Dec 19, 2022. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 0
Navigating To Machine Learning Environment
Daniel I Varzari edited this page Dec 6, 2022
·
2 revisions
The beautiful thing about Azure is that there is no installations required for the machine learning environment, all the data is either stored in the Data Lake directly, stored in spark tables in the Databricks tree menu for tables, or accessible through SQL commands through a connection to our PostgreSQL server.
- Launch Databricks from the resource group
- Navigate to repos where you wish to develop
- Before running a notebook attach a compute
- You can now run cells and expect an output
- Have Databricks launched
- Navigate to "compute" in tree
- Click create compute or configure an existing one
- Multi nodes are highly preferred
- Auto Scalability is what allows our compute resources to accept the massive parameters of our models
- Workers allow us to fall back on other computer resources if one get too overloaded 2-8 is sufficient for most applications
- Have Databricks launched
- Navigate to Machine Learning from tree
- View models - (here you can go into details/serving)
- View model histories
- Go to details for information about inputs and outputs of the endpoint
- Go to serving for more information about the url to connect to the model
- Have Databricks launched
- Go to source of model/navigate to a model notebook
- Click the flask
- View metrics and runtime details
- From these views you can reproduce model results
- We do not have to worry about re-running the model with new parameters so long as we are logging
- Logging with Sklearn is done through autologging
Wiki
Helpful Information
Helpful Guides