-
Notifications
You must be signed in to change notification settings - Fork 324
Initial_audit_changes #1436
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Initial_audit_changes #1436
Changes from all commits
05f79b4
afd202f
4bf3c81
63e50c7
c3d1f8f
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -37,7 +37,7 @@ pgml.train( | |
| `task` | `'regression'` | The objective of the experiment: `regression`, `classification` or `cluster` | | ||
| `relation_name` | `'public.search_logs'` | The Postgres table or view where the training data is stored or defined. | | ||
| `y_column_name` | `'clicked'` | The name of the label (aka "target" or "unknown") column in the training table. | | ||
| `algorithm` | `'xgboost'` | <p>The algorithm to train on the dataset, see the task specific pages for available algorithms:<br><a data-mention href="regression.md">regression.md</a></p><p><a data-mention href="classification.md">classification.md</a><br><a data-mention href="clustering.md">clustering.md</a></p> | | ||
| `algorithm` | `'xgboost'` | <p>The algorithm to train on the dataset, see the task specific pages for available algorithms:<br>[regression.md](regression.md "mention")</p><p>[classification.md](classification.md "mention")<br>[clustering.md](clustering.md "mention")</p> | | ||
| `hyperparams` | `{ "n_estimators": 25 }` | The hyperparameters to pass to the algorithm for training, JSON formatted. | | ||
| `search` | `grid` | If set, PostgresML will perform a hyperparameter search to find the best hyperparameters for the algorithm. See [hyperparameter-search.md](hyperparameter-search.md "mention") for details. | | ||
| `search_params` | `{ "n_estimators": [5, 10, 25, 100] }` | Search parameters used in the hyperparameter search, using the scikit-learn notation, JSON formatted. | | ||
|
@@ -63,7 +63,7 @@ This will create a "My Classification Project", copy the `pgml.digits` table int | |
|
||
When used for the first time in a project, `pgml.train()` function requires the `task` parameter, which can be either `regression` or `classification`. The task determines the relevant metrics and analysis performed on the data. All models trained within the project will refer to those metrics and analysis for benchmarking and deployment. | ||
|
||
The first time it's called, the function will also require a `relation_name` and `y_column_name`. The two arguments will be used to create the first snapshot of training and test data. By default, 25% of the data (specified by the `test_size` parameter) will be randomly sampled to measure the performance of the model after the `algorithm` has been trained on the 75% of the data. | ||
The first time it is called, the function will also require a `relation_name` and `y_column_name`. The two arguments will be used to create the first snapshot of training and test data. By default, 25% of the data (specified by the `test_size` parameter) will be randomly sampled to measure the performance of the model after the `algorithm` has been trained on the 75% of the data. | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Why do we prefer not using contractions? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I suggested it as a way to make it simpler for non-native English speakers to read. Another reason is for translation, but I figured you probably have no plans for that at this point. |
||
|
||
!!! tip | ||
|
||
|
Uh oh!
There was an error while loading. Please reload this page.