Skip to content

Commit 77985ee

Browse files
committed
Refine ML intro.
1 parent 109645f commit 77985ee

File tree

4 files changed

+52
-70
lines changed

4 files changed

+52
-70
lines changed

explore-analyze/machine-learning.md

+52-7
Original file line numberDiff line numberDiff line change
@@ -5,14 +5,59 @@ mapped_urls:
55
- https://www.elastic.co/guide/en/serverless/current/machine-learning.html
66
---
77

8-
# Machine learning
8+
# What is Elastic Machine Learning? [machine-learning-intro]
99

10-
% What needs to be done: Align serverless/stateful
10+
{{ml-cap}} features analyze your data and generate models for its patterns of behavior.
11+
The type of analysis that you choose depends on the questions or problems you want to address and the type of data you have available.
1112

12-
% Scope notes: include references to trained model autoscaling where appropriate
13+
## Unsupervised {{ml}} [machine-learning-unsupervised]
1314

14-
% Use migrated content from existing pages that map to this page:
15+
There are two types of analysis that can deduce the patterns and relationships within your data without training or intervention: *{{anomaly-detect}}* and *{{oldetection}}*.
1516

16-
% - [ ] ./raw-migrated-files/stack-docs/machine-learning/index.md
17-
% - [ ] ./raw-migrated-files/stack-docs/machine-learning/machine-learning-intro.md
18-
% - [ ] ./raw-migrated-files/docs-content/serverless/machine-learning.md
17+
[{{anomaly-detect-cap}}](machine-learning/anomaly-detection.md) requires time series data.
18+
It constructs a probability model and can run continuously to identify unusual events as they occur. The model evolves over time; you can use its insights to forecast future behavior.
19+
20+
[{{oldetection-cap}}](machine-learning/data-frame-analytics/ml-dfa-finding-outliers.md) does not require time series data.
21+
It is a type of {{dfanalytics}} that identifies unusual points in a data set by analyzing how close each data point is to others and the density of the cluster of points around it.
22+
It does not run continuously; it generates a copy of your data set where each data point is annotated with an {{olscore}}.
23+
The score indicates the extent to which a data point is an outlier compared to other data points.
24+
25+
## Supervised {{ml}} [machine-learning-supervised]
26+
27+
There are two types of {{dfanalytics}} that require training data sets: *{{classification}}* and *{{regression}}*.
28+
29+
In both cases, the result is a copy of your data set where each data point is annotated with predictions and a trained model, which you can deploy to make predictions for new data.
30+
For more information, refer to [Introduction to supervised learning](machine-learning/data-frame-analytics/ml-dfa-overview.md#ml-supervised-workflow).
31+
32+
[{{classification-cap}}](machine-learning/data-frame-analytics/ml-dfa-classification.md) learns relationships between your data points in order to predict discrete categorical values, such as whether a DNS request originates from a malicious or benign domain.
33+
34+
[{{regression-cap}}](machine-learning/data-frame-analytics/ml-dfa-regression.md) learns relationships between your data points in order to predict continuous numerical values, such as the response time for a web request.
35+
36+
## Feature availability by project type [machine-learning-serverless-availability]
37+
38+
The {{ml-features}} that are available vary by project type:
39+
40+
* {{es-serverless}} projects have trained models.
41+
* {{observability}} projects have {{anomaly-jobs}}.
42+
* {{elastic-sec}} projects have {{anomaly-jobs}}, {{dfanalytics-jobs}}, and trained models.
43+
44+
## Synchronize saved objects [machine-learning-synchronize-saved-objects]
45+
46+
Before you can view your {{ml}} {dfeeds}, jobs, and trained models in {{kib}}, they must have saved objects.
47+
For example, if you used APIs to create your jobs, wait for automatic synchronization or go to the **{{ml-app}}** page and click **Synchronize saved objects**.
48+
49+
## Export and import jobs [machine-learning-export-and-import-jobs]
50+
51+
You can export and import your {{ml}} job and {{dfeed}} configuration details on the **{{ml-app}}** page.
52+
For example, you can export jobs from your test environment and import them in your production environment.
53+
54+
The exported file contains configuration details; it does not contain the {{ml}} models.
55+
For {{anomaly-detect}}, you must import and run the job to build a model that is accurate for the new environment.
56+
For {{dfanalytics}}, trained models are portable; you can import the job then transfer the model to the new cluster.
57+
Refer to [Exporting and importing {{dfanalytics}} trained models](machine-learning/data-frame-analytics/ml-trained-models.md#export-import).
58+
59+
There are some additional actions that you must take before you can successfully import and run your jobs:
60+
61+
* The {{data-sources}} that are used by {{anomaly-detect}} {dfeeds} and {{dfanalytics}} source indices must exist; otherwise, the import fails.
62+
* If your {{anomaly-jobs}} use custom rules with filter lists, the filter lists must exist; otherwise, the import fails.
63+
* If your {{anomaly-jobs}} were associated with calendars, you must create the calendar in the new environment and add your imported jobs to the calendar.

raw-migrated-files/docs-content/serverless/machine-learning.md

-36
This file was deleted.

raw-migrated-files/stack-docs/machine-learning/index.md

-3
This file was deleted.

raw-migrated-files/stack-docs/machine-learning/machine-learning-intro.md

-24
This file was deleted.

0 commit comments

Comments
 (0)