Skip to content

Commit

Permalink
Merge branch 'master' into event-log/use-alerts-actions-clients
Browse files Browse the repository at this point in the history
* master: (111 commits)
  Remove flaky note from gauge tests (#73240)
  Convert functional vega tests to ts and unskip tests (#72238)
  [Graph] Unskip graph tests (#72291)
  Add default Elasticsearch credentials to docs (#72617)
  [APM] Read body from indicesStats in upload-telemetry-data (#72732)
  The directory in the command was missing the /generated directory and would cause all definitions to be regenerated in the wrong place. (#72766)
  [KP] use new ES client in SO service (#72289)
  [Security Solution][Exceptions] Prevents value list entries from co-existing with non value list entries (#72995)
  Return EUI CSS to Shareable Runtime (#72990)
  Removed useless karma test (#73190)
  [INGEST_MANAGER] Make package config name blank for endpoint on Package Config create (#73082)
  [Ingest Manager] Support DEGRADED state in fleet agent event (#73104)
  [Security Solution][Detections] Change detections breadcrumb title (#73059)
  [ML] Fixing unnecessary deleting job polling (#73087)
  [ML] Fixing recognizer wizard create job button (#73025)
  [Composable template] Preview composite template (#72598)
  [Uptime] Use manual intervals for ping histogram (#72928)
  [Security Solution][Endpoint] Task/policy save modal text change, remove duplicate policy details text (#73130)
  [Maps] fix tile layer attibution text and attribution link validation errors (#73160)
  skip ingest pipeline api tests
  ...
  • Loading branch information
gmmorris committed Jul 27, 2020
2 parents 9dba118 + 9c2056d commit 2d63a67
Show file tree
Hide file tree
Showing 884 changed files with 20,470 additions and 12,013 deletions.
2 changes: 1 addition & 1 deletion .github/CODEOWNERS
Validating CODEOWNERS rules …
Original file line number Diff line number Diff line change
Expand Up @@ -115,7 +115,6 @@
/src/dev/ @elastic/kibana-operations
/src/setup_node_env/ @elastic/kibana-operations
/src/optimize/ @elastic/kibana-operations
/src/es_archiver/ @elastic/kibana-operations
/packages/*eslint*/ @elastic/kibana-operations
/packages/*babel*/ @elastic/kibana-operations
/packages/kbn-dev-utils*/ @elastic/kibana-operations
Expand All @@ -124,6 +123,7 @@
/packages/kbn-pm/ @elastic/kibana-operations
/packages/kbn-test/ @elastic/kibana-operations
/packages/kbn-ui-shared-deps/ @elastic/kibana-operations
/packages/kbn-es-archiver/ @elastic/kibana-operations
/src/legacy/server/keystore/ @elastic/kibana-operations
/src/legacy/server/pid/ @elastic/kibana-operations
/src/legacy/server/sass/ @elastic/kibana-operations
Expand Down
6 changes: 5 additions & 1 deletion docs/developer/advanced/running-elasticsearch.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,10 @@ This will run a snapshot of {es} that is usually built nightly. Read more about
----
yarn es snapshot
----
By default, two users are added to Elasticsearch:

- A superuser with username: `elastic` and password: `changeme`, which can be used to log into Kibana with.
- A user with username: `kibana_system` and password `changeme`. This account is used by the Kibana server to authenticate itself to Elasticsearch, and to perform certain actions on behalf of the end user. These credentials should be specified in your kibana.yml as described in <<using-kibana-with-security>>

See all available options, like how to specify a specific license, with the `--help` flag.

Expand Down Expand Up @@ -115,4 +119,4 @@ PUT _cluster/settings
}
----

Follow the cross-cluster search instructions for setting up index patterns to search across clusters (<<management-cross-cluster-search>>).
Follow the cross-cluster search instructions for setting up index patterns to search across clusters (<<management-cross-cluster-search>>).
5 changes: 0 additions & 5 deletions docs/developer/architecture/code-exploration.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -186,11 +186,6 @@ WARNING: Missing README.
Replaces the legacy ui/share module for registering share context menus.
- {kib-repo}blob/{branch}/src/plugins/status_page[statusPage]
WARNING: Missing README.
- {kib-repo}blob/{branch}/src/plugins/telemetry/README.md[telemetry]
Telemetry allows Kibana features to have usage tracked in the wild. The general term "telemetry" refers to multiple things:
Expand Down
11 changes: 6 additions & 5 deletions docs/management/alerting/alerts-and-actions-intro.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -4,22 +4,23 @@

beta[]

The *Alerts and Actions* UI lets you <<alert-management, see and control all the alerts>> in a space, and provides tools to <<connector-management, create and manage connectors>> so that alerts can trigger actions like notification, indexing, and ticketing.
The *Alerts and Actions* UI lets you <<alert-management, see and control all the alerts>> in a space, and provides tools to <<connector-management, create and manage connectors>> so that alerts can trigger actions like notification, indexing, and ticketing.

To manage alerting and connectors, open the menu, then go to *Stack Management > {kib} > Alerts and Actions*.
To manage alerting and connectors, open the menu,
then go to *Stack Management > Alerts and Insights > Alerts and Actions*.

[role="screenshot"]
image:management/alerting/images/alerts-and-actions-ui.png[Example alert listing in the Alerts and Actions UI]

[NOTE]
============================================================================
Similar to dashboards, alerts and connectors reside in a <<xpack-spaces, space>>.
The *Alerts and Actions* UI only shows alerts and connectors for the current space.
The *Alerts and Actions* UI only shows alerts and connectors for the current space.
============================================================================

[NOTE]
============================================================================
{es} also offers alerting capabilities through Watcher, which
can be managed through the <<watcher-ui, Watcher UI>>. See
can be managed through the <<watcher-ui, Watcher UI>>. See
<<alerting-concepts-differences>> for more information.
============================================================================
============================================================================
Original file line number Diff line number Diff line change
Expand Up @@ -2,17 +2,16 @@
[[adding-policy-to-index]]
=== Adding a policy to an index

To add a lifecycle policy to an index and view the status for indices
managed by a policy, open the menu, then go to *Stack Management > {es} > Index Management*. This page lists your
{es} indices, which you can filter by lifecycle status and lifecycle phase.
To add a lifecycle policy to an index and view the status for indices
managed by a policy, open the menu, then go to *Stack Management > Data > Index Management*.
This page lists your
{es} indices, which you can filter by lifecycle status and lifecycle phase.

To add a policy, select the index name and then select *Manage Index > Add lifecycle policy*.
You’ll see the policy name, the phase the index is in, the current
action, and if any errors occurred performing that action.
You’ll see the policy name, the phase the index is in, the current
action, and if any errors occurred performing that action.

To remove a policy from an index, select *Manage Index > Remove lifecycle policy*.

[role="screenshot"]
image::images/index_management_add_policy.png[][UI for adding a policy to an index]


Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
[[example-using-index-lifecycle-policy]]
=== Tutorial: Use {ilm-init} to manage {filebeat} time-based indices

With {ilm} ({ilm-init}), you can create policies that perform actions automatically
With {ilm} ({ilm-init}), you can create policies that perform actions automatically
on indices as they age and grow. {ilm-init} policies help you to manage
performance, resilience, and retention of your data during its lifecycle. This tutorial shows
you how to use {kib}’s *Index Lifecycle Policies* to modify and create {ilm-init}
Expand Down Expand Up @@ -59,7 +59,7 @@ output as described in {filebeat-ref}/filebeat-getting-started.html[Getting Star
{filebeat} includes a default {ilm-init} policy that enables rollover. {ilm-init}
is enabled automatically if you’re using the default `filebeat.yml` and index template.

To view the default policy in {kib}, open the menu, go to * Stack Management > {es} > Index Lifecycle Policies*,
To view the default policy in {kib}, open the menu, go to *Stack Management > Data > Index Lifecycle Policies*,
search for _filebeat_, and choose the _filebeat-version_ policy.

This policy initiates the rollover action when the index size reaches 50GB or
Expand Down Expand Up @@ -114,7 +114,7 @@ If meeting a specific retention time period is most important, you can create a
custom policy. For this option, you will use {filebeat} daily indices without
rollover.

. To create a custom policy, open the menu, go to *Stack Management > {es} > Index Lifecycle Policies*, then click
. To create a custom policy, open the menu, go to *Stack Management > Data > Index Lifecycle Policies*, then click
*Create policy*.

. Activate the warm phase and configure it as follows:
Expand Down Expand Up @@ -156,7 +156,7 @@ image::images/tutorial-ilm-custom-policy.png["Modify the custom policy to add a
[role="screenshot"]
image::images/tutorial-ilm-delete-phase-creation.png["Delete phase"]

. To configure the index to use the new policy, open the menu, then go to *Stack Management > {es} > Index Lifecycle
. To configure the index to use the new policy, open the menu, then go to *Stack Management > Data > Index Lifecycle
Policies*.

.. Find your {ilm-init} policy.
Expand Down
4 changes: 2 additions & 2 deletions docs/management/ingest-pipelines/ingest-pipelines.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ pipelines that perform common transformations and
enrichments on your data. For example, you might remove a field,
rename an existing field, or set a new field.

You’ll find *Ingest Node Pipelines* in *Management > Elasticsearch*. With this feature, you can:
You’ll find *Ingest Node Pipelines* in *Stack Management > Ingest*. With this feature, you can:

* View a list of your pipelines and drill down into details.
* Create a pipeline that defines a series of tasks, known as processors.
Expand All @@ -23,7 +23,7 @@ image:management/ingest-pipelines/images/ingest-pipeline-list.png["Ingest node p
The minimum required permissions to access *Ingest Node Pipelines* are
the `manage_pipeline` and `cluster:monitor/nodes/info` cluster privileges.

You can add these privileges in *Management > Security > Roles*.
You can add these privileges in *Stack Management > Security > Roles*.

[role="screenshot"]
image:management/ingest-pipelines/images/ingest-pipeline-privileges.png["Privileges required for Ingest Node Pipelines"]
Expand Down
7 changes: 4 additions & 3 deletions docs/management/managing-beats.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,8 @@

include::{asciidoc-dir}/../../shared/discontinued.asciidoc[tag=cm-discontinued]

To use the Central Management UI, open the menu, go to *Stack Management > {beats} > Central Management*, then define and
To use {beats} Central Management UI, open the menu, go to *Stack Management > Ingest >
{beats} Central Management*, then define and
manage configurations in a central location in {kib} and quickly deploy
configuration changes to all {beats} running across your enterprise. For more
about central management, see the related {beats} documentation:
Expand All @@ -17,8 +18,8 @@ about central management, see the related {beats} documentation:
This feature requires an Elastic license that includes {beats} central
management.
Don't have a license? You can start a 30-day trial. Open the menu, go to
*Stack Management > Elasticsearch > License Management*. At the end of the trial
Don't have a license? You can start a 30-day trial. Open the menu,
go to *Stack Management > Stack > License Management*. At the end of the trial
period, you can purchase a subscription to keep using central management. For
more information, see https://www.elastic.co/subscriptions and
<<managing-licenses>>.
Expand Down
2 changes: 1 addition & 1 deletion docs/management/managing-ccr.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ remote clusters on a local cluster. {ref}/xpack-ccr.html[Cross-cluster replicati
is commonly used to provide remote backups for disaster recovery and for
geo-proximite copies of data.

To get started, open the menu, then go to *Stack Management > Elasticsearch > Cross-Cluster Replication*.
To get started, open the menu, then go to *Stack Management > Data > Cross-Cluster Replication*.

[role="screenshot"]
image::images/cross-cluster-replication-list-view.png[][Cross-cluster replication list view]
Expand Down
20 changes: 10 additions & 10 deletions docs/management/managing-indices.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ the amount of bookkeeping when working with indices. Instead of manually
setting up your indices, you can create them automatically from a template,
ensuring that your settings, mappings, and aliases are consistently defined.

To manage your indices, open the menu, then go to *Stack Management > {es} > Index Management*.
To manage your indices, open the menu, then go to *Stack Management > Data > Index Management*.

[role="screenshot"]
image::images/management_index_labels.png[Index Management UI]
Expand Down Expand Up @@ -130,17 +130,17 @@ Alternatively, you can click the *Load JSON* link and define the mapping as JSON

[source,js]
----------------------------------
{
{
"properties": {
"geo": {
"properties": {
"coordinates": {
"type": "geo_point"
}
}
}
}
}
"properties": {
"coordinates": {
"type": "geo_point"
}
}
}
}
}
----------------------------------

You can create additional mapping configurations in the *Dynamic templates* and
Expand Down
2 changes: 1 addition & 1 deletion docs/management/managing-licenses.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ with no expiration date. For the full list of features, refer to

If you want to try out the full set of features, you can activate a free 30-day
trial. To view the status of your license, start a trial, or install a new
license, open the menu, then go to *Stack Management > {es} > License Management*.
license, open the menu, then go to *Stack Management > Stack > License Management*.

NOTE: You can start a trial only if your cluster has not already activated a
trial license for the current major product version. For example, if you have
Expand Down
2 changes: 1 addition & 1 deletion docs/management/managing-remote-clusters.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ connection from your cluster to other clusters. This functionality is
required for {ref}/xpack-ccr.html[cross-cluster replication] and
{ref}/modules-cross-cluster-search.html[cross-cluster search].

To get started, open the menu, then go to *Stack Management > {es} > Remote Clusters*.
To get started, open the menu, then go to *Stack Management > Data > Remote Clusters*.

[role="screenshot"]
image::images/remote-clusters-list-view.png[Remote Clusters list view, including Add a remote cluster button]
Expand Down
11 changes: 7 additions & 4 deletions docs/management/rollups/create_and_manage_rollups.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ by an index pattern, and then rolls it into a new index. Rollup indices are a go
compactly store months or years of historical
data for use in visualizations and reports.

To get started, open the menu, then go to *Stack Management > {es} > Rollup Jobs*. With this UI,
To get started, open the menu, then go to *Stack Management > Data > Rollup Jobs*. With this UI,
you can:

* <<create-and-manage-rollup-job, Create a rollup job>>
Expand Down Expand Up @@ -130,8 +130,9 @@ Your next step is to visualize your rolled up data in a vertical bar chart.
Most visualizations support rolled up data, with the exception of Timelion and Vega visualizations.


. Create the rollup index pattern in *Management > Index Patterns* so you can
select your rolled up data for visualizations. Click *Create index pattern*, and select *Rollup index pattern* from the dropdown.
. Go to *Stack Management > {kib} > Index Patterns*.

. Click *Create index pattern*, and select *Rollup index pattern* from the dropdown.
+
[role="screenshot"]
image::images/management-rollup-index-pattern.png[][Create rollup index pattern]
Expand All @@ -144,7 +145,9 @@ is `rollup_logstash,kibana_sample_data_logs`. In this index pattern, `rollup_log
matches the rolled up index pattern and `kibana_sample_data_logs` matches the index
pattern for raw data.

. Go to *Visualize* and create a vertical bar chart. Choose `rollup_logstash,kibana_sample_data_logs`
. Go to *Visualize* and create a vertical bar chart.

. Choose `rollup_logstash,kibana_sample_data_logs`
as your source to see both the raw and rolled up data.
+
[role="screenshot"]
Expand Down
4 changes: 2 additions & 2 deletions docs/management/snapshot-restore/index.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ Snapshots are important because they provide a copy of your data in case
something goes wrong. If you need to roll back to an older version of your data,
you can restore a snapshot from the repository.

To get started, open the menu, then go to *Stack Management > {es} > Snapshot and Restore*.
To get started, open the menu, then go to *Stack Management > Data > Snapshot and Restore*.
With this UI, you can:

* Register a repository for storing your snapshots
Expand Down Expand Up @@ -191,7 +191,7 @@ your master and data nodes. You can do this in one of two ways:
Use *Snapshot and Restore* to register the repository where your snapshots
will live.

. Open the menu, then go to *Stack Management > {es} > Snapshot and Restore*.
. Open the menu, then go to *Stack Management > Data > Snapshot and Restore*.
. Click *Register a repository* in either the introductory message or *Repository view*.
. Enter a name for your repository, for example, `my_backup`.
. Select *Shared file system*.
Expand Down
42 changes: 21 additions & 21 deletions docs/management/upgrade-assistant/index.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -2,50 +2,50 @@
[[upgrade-assistant]]
== Upgrade Assistant

The Upgrade Assistant helps you prepare for your upgrade to the next major {es} version.
For example, if you are using 6.8, the Upgrade Assistant helps you to upgrade to 7.0.
To access the assistant, open the menu, then go to *Stack Management > {es} > Upgrade Assistant*.
The Upgrade Assistant helps you prepare for your upgrade to the next major {es} version.
For example, if you are using 6.8, the Upgrade Assistant helps you to upgrade to 7.0.
To access the assistant, open the menu, then go to *Stack Management > Stack > Upgrade Assistant*.

The assistant identifies the deprecated settings in your cluster and indices
and guides you through the process of resolving issues, including reindexing.
The assistant identifies the deprecated settings in your cluster and indices
and guides you through the process of resolving issues, including reindexing.

Before you upgrade, make sure that you are using the latest released minor
version of {es} to see the most up-to-date deprecation issues.
Before you upgrade, make sure that you are using the latest released minor
version of {es} to see the most up-to-date deprecation issues.
For example, if you want to upgrade to to 7.0, make sure that you are using 6.8.

[float]
=== Reindexing

The *Indices* page lists the indices that are incompatible with the next
The *Indices* page lists the indices that are incompatible with the next
major version of {es}. You can initiate a reindex to resolve the issues.

[role="screenshot"]
image::images/management-upgrade-assistant-9.0.png[]

For a preview of how the data will change during the reindex, select the
index name. A warning appears if the index requires destructive changes.
Back up your index, then proceed with the reindex by accepting each breaking change.
For a preview of how the data will change during the reindex, select the
index name. A warning appears if the index requires destructive changes.
Back up your index, then proceed with the reindex by accepting each breaking change.

You can follow the progress as the Upgrade Assistant makes the index read-only,
creates a new index, reindexes the documents, and creates an alias that points
from the old index to the new one.
You can follow the progress as the Upgrade Assistant makes the index read-only,
creates a new index, reindexes the documents, and creates an alias that points
from the old index to the new one.

If the reindexing fails or is cancelled, the changes are rolled back, the
new index is deleted, and the original index becomes writable. An error
If the reindexing fails or is cancelled, the changes are rolled back, the
new index is deleted, and the original index becomes writable. An error
message explains the reason for the failure.

You can reindex multiple indices at a time, but keep an eye on the
{es} metrics, including CPU usage, memory pressure, and disk usage. If a
metric is so high it affects query performance, cancel the reindex and
You can reindex multiple indices at a time, but keep an eye on the
{es} metrics, including CPU usage, memory pressure, and disk usage. If a
metric is so high it affects query performance, cancel the reindex and
continue by reindexing fewer indices at a time.

Additional considerations:

* If you use {alert-features}, when you reindex the internal indices
(`.watches`), the {watcher} process pauses and no alerts are triggered.
(`.watches`), the {watcher} process pauses and no alerts are triggered.

* If you use {ml-features}, when you reindex the internal indices (`.ml-state`),
the {ml} jobs pause and models are not trained or updated.
the {ml} jobs pause and models are not trained or updated.

* If you use {security-features}, before you reindex the internal indices
(`.security*`), it is a good idea to create a temporary superuser account in the
Expand Down
6 changes: 2 additions & 4 deletions docs/management/watcher-ui/index.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,8 @@ Watches are helpful for analyzing mission-critical and business-critical
streaming data. For example, you might watch application logs for performance
outages or audit access logs for security threats.

To get started with the Watcher UI, open then menu, then go to *Stack Management > {es} > Watcher*.
To get started with the Watcher UI, open then menu,
then go to *Stack Management > Alerts and Insights > Watcher*.
With this UI, you can:

* <<watcher-create-threshold-alert, Create a simple threshold watch>>
Expand Down Expand Up @@ -238,6 +239,3 @@ Refer to these examples for creating an advanced watch:

* {ref}/watch-cluster-status.html[Watch the status of an {es} cluster]
* {ref}/watching-meetup-data.html[Watch event data]



Loading

0 comments on commit 2d63a67

Please sign in to comment.