From b59760925481bbce9d76bb095884d248df9165ca Mon Sep 17 00:00:00 2001 From: ditaccms-bot Date: Fri, 2 Aug 2024 14:42:25 +0000 Subject: [PATCH] Update from SAP DITA CMS (squashed): commit b3fd805af6a24753236db5bfa319a1566df2093f Author: REDACTED Date: Wed Jul 31 10:25:09 2024 +0000 Update from SAP DITA CMS 2024-07-31 10:25:09 Project: dita-all/urr1719873541902 Project map: af2fcb3e6dd448f3af3c0ff9c70daaf9.ditamap Output: loioc25299a38b6448f889a43b42c9e5897d Language: en-US Builddable map: 678695d903b546e5947af69e56ed42b8.ditamap commit f80584572a1530ab85e34eb6b2d8874167fe30ab Author: REDACTED Date: Wed Jul 31 10:24:01 2024 +0000 Update from SAP DITA CMS 2024-07-31 10:24:01 Project: dita-all/urr1719873541902 Project map: af2fcb3e6dd448f3af3c0ff9c70daaf9.ditamap Output: loiob8faae83b519439fb4ea9d0eb1a5f26e Language: en-US Builddable map: 4e1c1e1d5d1947f5875e93e7597c4f4c.ditamap commit ad98d8b83d5d774f911e9423e021951428c99666 Author: REDACTED Date: Wed Jul 31 10:21:50 2024 +0000 Update from SAP DITA CMS 2024-07-31 10:21:50 Project: dita-all/urr1719873541902 Project map: af2fcb3e6dd448f3af3c0ff9c70daaf9.ditamap Output: loiod3d776bb52294a17b48298443a286f55 Language: en-US Builddable map: 89ab8c0ed18c432d8fb87551823e7de7.ditamap ################################################## [Remaining squash message was removed before commit...] --- .../add-a-source-7b50e8e.md | 5 + .../running-a-flow-5b591d4.md | 3 + ...thout-primary-key-as-the-source-2267a9f.md | 22 +++ ...g-confluent-kafka-as-the-target-74b3c95.md | 2 +- ...d-accessing-data-in-the-catalog-1047825.md | 11 ++ .../add-a-dimension-4caf098.md | 1 + .../add-a-prefix-or-a-suffix-0373c60.md | 30 ++++ .../index.md | 2 + .../searching-and-browsing-2dfea3a.md | 5 + .../viewing-object-data-b338e4a.md | 10 +- .../create-a-space-bbd41b8.md | 6 +- ...er-58d4b24.md => create-a-user-58d4b24.md} | 2 +- ...ers-3ceb94c.md => delete-users-3ceb94c.md} | 8 +- ...a-custom-saml-identity-provider-9b26536.md | 2 +- ...ers-e227d3c.md => export-users-e227d3c.md} | 2 +- ...rt-or-modify-users-from-a-file-b2698da.md} | 6 +- .../managing-sap-datasphere-users-4fb82cb.md | 10 +- ...d-privileges-by-app-and-feature-2d8b7d0.md | 2 +- ...s-delivered-with-sap-datasphere-a50a51d.md | 2 +- ...=> update-user-email-addresses-0889208.md} | 6 +- ...lyze-monitoring-data-in-a-space-9cd0691.md | 4 +- ...s-user-to-debug-database-issues-c28145b.md | 2 +- ...base-operations-with-audit-logs-110404a.md | 2 +- .../monitoring-sap-datasphere-28910cd.md | 14 +- .../configure-cloud-connector-f289920.md | 12 ++ .../prepare-connectivity-to-oracle-9fca7c4.md | 9 +- ...nnectivity-to-sap-s-4hana-cloud-abb159e.md | 20 +-- ...ng-connectivity-for-connections-bffbd58.md | 4 +- ...drivers-required-for-data-flows-b9b5579.md | 4 +- ...ing-data-from-sap-s-4hana-cloud-ef2b223.md | 94 +++++++++++ .../administration-apps-and-tools-c6dd052.md | 2 +- docs/Administering/index.md | 11 +- ...ovisioning-agent-status-changes-85790bb.md | 12 +- .../request-help-from-sap-support-831a977.md | 47 +++--- ...atistics-for-your-remote-tables-e4120bb.md | 2 +- .../monitoring-local-tables-3740461.md | 8 + ...ple-storage-service-connections-a7b660a.md | 7 +- ...ributed-file-system-connections-f9c3356.md | 2 +- ...ata-lake-store-gen2-connections-cd06b3c.md | 2 +- .../oracle-connections-c73ae06.md | 151 ++++++++++++++++-- ...oud-data-lake-files-connections-356e41e.md | 2 +- ...e-relational-engine-connections-40763e2.md | 2 +- .../sap-hana-connections-e6b63f1.md | 8 +- .../sap-s-4hana-cloud-connections-a98e5ff.md | 111 ++++++++++++- .../sap-successfactors-connections-39df020.md | 2 + ...-sap-cloud-transport-management-0538398.md | 4 +- .../lock-or-unlock-your-space-c05b6a6.md | 2 +- 47 files changed, 554 insertions(+), 123 deletions(-) create mode 100644 docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/using-an-object-without-primary-key-as-the-source-2267a9f.md create mode 100644 docs/Acquiring-Preparing-Modeling-Data/Modeling-Data-in-the-Data-Builder/add-a-prefix-or-a-suffix-0373c60.md rename docs/Administering/Managing-Users-and-Roles/{creating-a-new-user-58d4b24.md => create-a-user-58d4b24.md} (99%) rename docs/Administering/Managing-Users-and-Roles/{deleting-users-3ceb94c.md => delete-users-3ceb94c.md} (55%) rename docs/Administering/Managing-Users-and-Roles/{exporting-users-e227d3c.md => export-users-e227d3c.md} (99%) rename docs/Administering/Managing-Users-and-Roles/{importing-or-modifying-users-from-a-file-b2698da.md => import-or-modify-users-from-a-file-b2698da.md} (96%) rename docs/Administering/Managing-Users-and-Roles/{updating-user-email-addresses-0889208.md => update-user-email-addresses-0889208.md} (70%) create mode 100644 docs/Administering/Preparing-Connectivity/using-abap-sql-services-for-accessing-data-from-sap-s-4hana-cloud-ef2b223.md diff --git a/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/add-a-source-7b50e8e.md b/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/add-a-source-7b50e8e.md index b394a2d..1fb9388 100644 --- a/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/add-a-source-7b50e8e.md +++ b/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/add-a-source-7b50e8e.md @@ -71,6 +71,11 @@ Add a source to read data from. You can add multiple sources and combine them to \[read-only\] Provide information to identify the source table. + + > ### Note: + > Business names are not supported for columns in data flow editor. + + diff --git a/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/running-a-flow-5b591d4.md b/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/running-a-flow-5b591d4.md index 3f3bba9..0b3f96e 100644 --- a/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/running-a-flow-5b591d4.md +++ b/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/running-a-flow-5b591d4.md @@ -12,6 +12,9 @@ To run a flow, you have 3 main options depending on your flow type: - Click *Schedule* to run your data flow or your transformation flow at later time, or on regular basis \(this is not available for *Replication Flow*\). - Add the flow into a task chain. This is only valid for a data flow. For more information, see [Creating a Task Chain](creating-a-task-chain-d1afbc2.md). +> ### Caution: +> A data flow uses a technical user during run. So, if a source view or an underlying view of the source view uses the SESSION\_CONTEXT function, incorrect data is read from the view during the execution of the data flow. This results in incorrect data or no data in the output. + diff --git a/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/using-an-object-without-primary-key-as-the-source-2267a9f.md b/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/using-an-object-without-primary-key-as-the-source-2267a9f.md new file mode 100644 index 0000000..48a81b1 --- /dev/null +++ b/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/using-an-object-without-primary-key-as-the-source-2267a9f.md @@ -0,0 +1,22 @@ + + +# Using an Object Without Primary Key As the Source + +If certain conditions are met, you can use objects that do not have a primary key as the source for a replication flow. + +This applies to the following types of **source objects**: + +- CDS views + +- ODP artifacts + + +The objects must have load type *Initial Only*. \(Delta loading is not supported.\) + +**Existing tables** can be used as **targets** if they have a column \_\_load\_package\_id \(as explained below\). **Local tables** can only be used as targets if delta capturing is **not** enabled. + +For some targets, the system automatically adds a **technical target column**. For HANA and HANA Cloud targets, this column is called "\_\_load\_package\_id" and has data type binary\(\>=256\). For other targets, it is called "\_\_load\_record\_id" and has data type string\(44\). This column provides a unique identifier for each record, which serves as a replacement for the primary key in duplicate handling and for other technical purposes. On the Projections tab, this column is always shown at the end and in read only mode. You can't rename it, remove or modify its data type, or reorder it. + +> ### Note: +> Multiple replications \(for example due to restarting or undeploying and redeploying a replication flow\) result in different \_\_load\_record\_id values for the same source record. + diff --git a/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/using-confluent-kafka-as-the-target-74b3c95.md b/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/using-confluent-kafka-as-the-target-74b3c95.md index d6d955a..c694912 100644 --- a/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/using-confluent-kafka-as-the-target-74b3c95.md +++ b/docs/Acquiring-Preparing-Modeling-Data/Acquiring-and-Preparing-Data-in-the-Data-Builder/using-confluent-kafka-as-the-target-74b3c95.md @@ -171,7 +171,7 @@ For more information, see [Using Confluent Kafka As the Target](using-confluent- The **target container** is automatically set to "/" because Confluent Kafka does not have a superordinate container layer. > ### Note: -> If the Kafka cluster is behind an SAP Cloud Connector \(SCC\), the Kafka cluster and the SCC must be configured such that the broker addresses advertised by the cluster match the virtual hosts maintained for the brokers in the SCC. The simplest solution is to use the same value for virtual and internal hosts in the SCC and to maintain no dedicated advertised listeners for the Kafka brokers. If advertised listeners are maintained, these must be used as virtual hosts in SCC and as broker addresses in the connection definition. +> If the Kafka cluster is behind an SAP Cloud Connector \(SCC\), the Kafka cluster and the SCC must be configured such that the broker addresses advertised by the cluster match the virtual hosts maintained for the brokers in the SCC. The simplest solution is to use the same value for virtual and internal hosts in the SCC and to maintain no dedicated advertised listeners for the Kafka brokers. When adding additional brokers to a cluster to scale up, the SCC mapping must be adapted to include the additional brokers. Each record from the source system is transferred into a single **message** in the target topic. The key of the messages is the combination of all primary key values of the record concatenated by "\_". diff --git a/docs/Acquiring-Preparing-Modeling-Data/Creating-Finding-Sharing-Objects/finding-and-accessing-data-in-the-catalog-1047825.md b/docs/Acquiring-Preparing-Modeling-Data/Creating-Finding-Sharing-Objects/finding-and-accessing-data-in-the-catalog-1047825.md index 81ea574..a111dc9 100644 --- a/docs/Acquiring-Preparing-Modeling-Data/Creating-Finding-Sharing-Objects/finding-and-accessing-data-in-the-catalog-1047825.md +++ b/docs/Acquiring-Preparing-Modeling-Data/Creating-Finding-Sharing-Objects/finding-and-accessing-data-in-the-catalog-1047825.md @@ -264,6 +264,17 @@ You can change the view to see different information. - \(Display as List\) - \(Display as Table\) + + + + +## Search for Third-Party Data + +You can use the Datarade Data Finder website to search and request additional third-party data products from various data providers. Access the Datarade website from the catalog by choosing the *Data Products* tab or the *Data Providers* tab, and then clicking the *Request Data from Datarade* button. + +> ### Note: +> Clicking the *Request Data from Datarade* button takes you to the Datarade website. This site is external to our product. + **Related Information** diff --git a/docs/Acquiring-Preparing-Modeling-Data/Modeling-Data-in-the-Data-Builder/add-a-dimension-4caf098.md b/docs/Acquiring-Preparing-Modeling-Data/Modeling-Data-in-the-Data-Builder/add-a-dimension-4caf098.md index b6ebf71..45ff49b 100644 --- a/docs/Acquiring-Preparing-Modeling-Data/Modeling-Data-in-the-Data-Builder/add-a-dimension-4caf098.md +++ b/docs/Acquiring-Preparing-Modeling-Data/Modeling-Data-in-the-Data-Builder/add-a-dimension-4caf098.md @@ -46,5 +46,6 @@ There are different ways to add dimensions: - You can change the alias of this item in the properties panel. The alias is the name that is shown in the story in SAP Analytics Cloud. - You can add or deselect associated dimensions. - You can add or deselect attributes. + - You can specify a custom prefix or suffix for the technical name of attributes. For more information, see [Add a Prefix or a Suffix](add-a-prefix-or-a-suffix-0373c60.md). diff --git a/docs/Acquiring-Preparing-Modeling-Data/Modeling-Data-in-the-Data-Builder/add-a-prefix-or-a-suffix-0373c60.md b/docs/Acquiring-Preparing-Modeling-Data/Modeling-Data-in-the-Data-Builder/add-a-prefix-or-a-suffix-0373c60.md new file mode 100644 index 0000000..7addb89 --- /dev/null +++ b/docs/Acquiring-Preparing-Modeling-Data/Modeling-Data-in-the-Data-Builder/add-a-prefix-or-a-suffix-0373c60.md @@ -0,0 +1,30 @@ + + + + +# Add a Prefix or a Suffix + +You can specify a custom prefix or suffix for the technical name of attributes. + + + +## Context + +The suffix/prefix is a dimension property that is added to all attributes of that dimension. This helps you to identify where an attribute originates from when your analytic model contains a dimension multiple times, as the attribute name is be taken from the source dimension and would be the same in each dimension. + +> ### Example: +> An analytic model contains the dimension *Employee* three times, as *Product Manager*, *Responsible* and *Changed By*. You can add the prefix PM for *Employee* in *Product Manager*, the prefix RESP for *Employee* in *Responsible*, and CB for the *Employee* in *Changed By*. + + + +## Procedure + +1. Click on a dimension of the canvas to show its properties in the side panel. + +2. Choose Edit to edit the prefix or suffix for the technical name of attributes. Note that prefixes cannot start with an underscore. + +3. You can reset the technical names of attributes. + + When you select *Reset all attributes to the original source names*, you can reset the technical names. When an attribute's technical name is renamed with a counter because of conflicts, this function removes that counter by resetting the technical name to match the source. This does not affect prefixes and suffixes. + + diff --git a/docs/Acquiring-Preparing-Modeling-Data/index.md b/docs/Acquiring-Preparing-Modeling-Data/index.md index 57c74dc..6e69970 100644 --- a/docs/Acquiring-Preparing-Modeling-Data/index.md +++ b/docs/Acquiring-Preparing-Modeling-Data/index.md @@ -80,6 +80,7 @@ - [Configure Your Replication Flow](Acquiring-and-Preparing-Data-in-the-Data-Builder/configure-your-replication-flow-3f5ba0c.md) - [Define Filters](Acquiring-and-Preparing-Data-in-the-Data-Builder/define-filters-5a6ef36.md) - [Define Mapping](Acquiring-and-Preparing-Data-in-the-Data-Builder/define-mapping-2c7948f.md) + - [Using an Object Without Primary Key As the Source](Acquiring-and-Preparing-Data-in-the-Data-Builder/using-an-object-without-primary-key-as-the-source-2267a9f.md) - [Using a Cloud Storage Provider As the Target](Acquiring-and-Preparing-Data-in-the-Data-Builder/using-a-cloud-storage-provider-as-the-target-43d93a2.md) - [Using Google BigQuery As the Target](Acquiring-and-Preparing-Data-in-the-Data-Builder/using-google-bigquery-as-the-target-56d4472.md) - [Using Apache Kafka As the Target](Acquiring-and-Preparing-Data-in-the-Data-Builder/using-apache-kafka-as-the-target-6df55db.md) @@ -175,6 +176,7 @@ - [Create an Analytic Model Directly From a View or Table](Modeling-Data-in-the-Data-Builder/create-an-analytic-model-directly-from-a-view-or-table-1c674aa.md) - [Add a Source](Modeling-Data-in-the-Data-Builder/add-a-source-27075ee.md) - [Add a Dimension](Modeling-Data-in-the-Data-Builder/add-a-dimension-4caf098.md) + - [Add a Prefix or a Suffix](Modeling-Data-in-the-Data-Builder/add-a-prefix-or-a-suffix-0373c60.md) - [Add Measures](Modeling-Data-in-the-Data-Builder/add-measures-e4cc3e8.md) - [Create a Calculated Measure](Modeling-Data-in-the-Data-Builder/create-a-calculated-measure-cf6bd08.md) - [Create a Restricted Measure](Modeling-Data-in-the-Data-Builder/create-a-restricted-measure-bfb43dd.md) diff --git a/docs/Acquiring-Preparing-Modeling-Data/searching-and-browsing-2dfea3a.md b/docs/Acquiring-Preparing-Modeling-Data/searching-and-browsing-2dfea3a.md index f92789e..8eed2f5 100644 --- a/docs/Acquiring-Preparing-Modeling-Data/searching-and-browsing-2dfea3a.md +++ b/docs/Acquiring-Preparing-Modeling-Data/searching-and-browsing-2dfea3a.md @@ -18,3 +18,8 @@ For more information on the available filters, see [Overview of Filters](overvie By clicking an entry in the search result list, you will be directly forwarded to the corresponding data product page or data provider profile. There, you can evaluate the data product or data provider and eventually bookmark them. For more information, see [Using Bookmarks](using-bookmarks-aec68bb.md). +You can use the Datarade data finder website to search and request additional third-party data products from various data providers. Access Datarade by clicking the *Request Data from Datarade* button. + +> ### Note: +> Clicking the *Request Data from Datarade* button takes you to the Datarade website. This site is external to our product. + diff --git a/docs/Acquiring-Preparing-Modeling-Data/viewing-object-data-b338e4a.md b/docs/Acquiring-Preparing-Modeling-Data/viewing-object-data-b338e4a.md index 7cb2e2c..c86efb4 100644 --- a/docs/Acquiring-Preparing-Modeling-Data/viewing-object-data-b338e4a.md +++ b/docs/Acquiring-Preparing-Modeling-Data/viewing-object-data-b338e4a.md @@ -43,8 +43,6 @@ You can view data in the table, graphical view, SQL view, E/R model, analytic mo > - The *Data Viewer* will time out and display an error on the *Errors* tab if it is unable to load data in under three minutes. If you experience timeouts or slow performance and some or all of the data you want to view is not replicated to SAP Datasphere, consider replicating it \(see [Replicate Remote Table Data](Acquiring-and-Preparing-Data-in-the-Data-Builder/replicate-remote-table-data-7e258a7.md)\). > - If a column has a *HANA Numeric* data type its data cannot be shown. Consider, changing the data type to one supported by SAP Datasphere \(see [Column Data Types](Acquiring-and-Preparing-Data-in-the-Data-Builder/column-data-types-7b1dc6e.md)\). > - If a column has *Binary*, *Large Binary*, *ST\_Point*, and *ST\_Geometry* set as data type, the data viewer is unavailable and records are shown as *Cannot be shown*. If no data can be found, the data viewer shows it as *NULL*. - > - You cannot view data in an SQL view if any of its sources is shared from another space and has an input parameter. - > > - When viewing data in a persisted view, the data is read from the view's persistency table, unless the view has a status of *Changes to Deploy*, in which case the view is run and the results displayed. 4. The following tools are available in the panel toolbar: @@ -218,8 +216,14 @@ You can review the effects of any data access controls you apply to a view by ch The data viewer updates to show only the records visible to the selected user. > ### Note: - > *View as User* is only available in the graphic view, SQL view, and analytic model editors, and can only simulate data access controls that are applied in the current space and have a status of *Deployed*. If your view has one or more sources shared from other spaces, then a warning is shown to indicate that you may not see exactly the same records as the user you are impersonating. + > *View as User* is only available in the graphic view, SQL view, and analytic model editors, and can only simulate data access controls that are applied in the current space and which have a status of *Deployed*. In addition, if the view: + > + > - Is shared from another space then *View as User* is not supported. + > - Has one or more sources shared from other spaces, then a warning is shown to indicate that you may not see exactly the same records as the user you are impersonating. + > - Is secured by data access controls created by importing analysis authorizations from SAP BW or SAP BW/4HANA, then *View as User* is not supported. + +For more information about data access controls, see [Securing Data with Data Access Controls](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/a032e51c730147c7a1fcac125b4cfe14.html "Data access controls allow you to apply row-level security to your objects. When a data access control is applied to a data layer view or a business layer object, any user viewing its data will see only the rows for which they are authorized, based on the specified criteria.") :arrow_upper_right:. diff --git a/docs/Administering/Creating-Spaces-and-Allocating-Storage/create-a-space-bbd41b8.md b/docs/Administering/Creating-Spaces-and-Allocating-Storage/create-a-space-bbd41b8.md index 60d8213..84386cb 100644 --- a/docs/Administering/Creating-Spaces-and-Allocating-Storage/create-a-space-bbd41b8.md +++ b/docs/Administering/Creating-Spaces-and-Allocating-Storage/create-a-space-bbd41b8.md @@ -243,7 +243,7 @@ Create a space, allocate storage, and set the space priority and statement limit > ### Note: > A user with the DW Administrator role only cannot see the *Time Tables and Dimensions* area. - - *Auditing*/*Space Audit Settings* - Use the properties in the *Space Audit Settings* section to enable audit logging for the space. See [Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who did what and when in the database.") :arrow_upper_right:. + - *Auditing*/*Space Audit Settings* - Use the properties in the *Space Audit Settings* section to enable audit logging for the space. See [Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who performed which action at which point in time.") :arrow_upper_right:. 7. Click *Deploy* to deploy your space to the database. @@ -252,8 +252,8 @@ Create a space, allocate storage, and set the space priority and statement limit - Add your space to an existing scoped role \(see [Add Spaces to a Scoped Role](../Managing-Users-and-Roles/create-a-scoped-role-to-assign-privileges-to-users-in-spaces-b5c4e0b.md#loiob5c4e0b6c462414783ebbfc053815521__section_pr1_5pj_zyb)\). - Create a scoped role and add your space and at least one user to the scoped role \(see [Create a Scoped Role](../Managing-Users-and-Roles/create-a-scoped-role-to-assign-privileges-to-users-in-spaces-b5c4e0b.md#loiob5c4e0b6c462414783ebbfc053815521__section_z4m_mpj_zyb)\). - For more information, see [Create a Scoped Role to Assign Privileges to Users in Spaces](../Managing-Users-and-Roles/create-a-scoped-role-to-assign-privileges-to-users-in-spaces-b5c4e0b.md)\). + For more information, see [Create a Scoped Role to Assign Privileges to Users in Spaces](../Managing-Users-and-Roles/create-a-scoped-role-to-assign-privileges-to-users-in-spaces-b5c4e0b.md). - All users assigned to the space via the scoped roles are automatically displayed in the *Users* area of the space page. In this area, you can add or remove users to/from scoped roles for your space \(see [Control User Access to Your Space](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/9d59fe511ae644d98384897443054c16.html "You can assign users to your space and manage them.") :arrow_upper_right:. Either an administrator or a user with space administrator privileges can do so. + All users assigned to the space via the scoped roles are automatically displayed in the *Users* area of the space page. In this area, you can add or remove users to/from scoped roles for your space \(see [Control User Access to Your Space](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/9d59fe511ae644d98384897443054c16.html "You can assign users to your space and manage them.") :arrow_upper_right:\). Either an administrator or a user with space administrator privileges can do so. diff --git a/docs/Administering/Managing-Users-and-Roles/creating-a-new-user-58d4b24.md b/docs/Administering/Managing-Users-and-Roles/create-a-user-58d4b24.md similarity index 99% rename from docs/Administering/Managing-Users-and-Roles/creating-a-new-user-58d4b24.md rename to docs/Administering/Managing-Users-and-Roles/create-a-user-58d4b24.md index 7b5e4bb..1679e3d 100644 --- a/docs/Administering/Managing-Users-and-Roles/creating-a-new-user-58d4b24.md +++ b/docs/Administering/Managing-Users-and-Roles/create-a-user-58d4b24.md @@ -2,7 +2,7 @@ -# Creating a New User +# Create a User You can create individual users in SAP Datasphere. diff --git a/docs/Administering/Managing-Users-and-Roles/deleting-users-3ceb94c.md b/docs/Administering/Managing-Users-and-Roles/delete-users-3ceb94c.md similarity index 55% rename from docs/Administering/Managing-Users-and-Roles/deleting-users-3ceb94c.md rename to docs/Administering/Managing-Users-and-Roles/delete-users-3ceb94c.md index a23f8fb..efee6c3 100644 --- a/docs/Administering/Managing-Users-and-Roles/deleting-users-3ceb94c.md +++ b/docs/Administering/Managing-Users-and-Roles/delete-users-3ceb94c.md @@ -2,7 +2,7 @@ -# Deleting Users +# Delete Users You can delete users. @@ -24,9 +24,9 @@ You can delete users. **Related Information** -[Creating a New User](creating-a-new-user-58d4b24.md "You can create individual users in SAP Datasphere.") +[Create a User](create-a-user-58d4b24.md "You can create individual users in SAP Datasphere.") -[Importing or Modifying Users from a File](importing-or-modifying-users-from-a-file-b2698da.md "You can create new users or batch-update existing users by importing user data that you have saved in a CSV file.") +[Import or Modify Users from a File](import-or-modify-users-from-a-file-b2698da.md "You can create users or batch-update existing users by importing user data that you have saved in a CSV file.") -[Updating User Email Addresses](updating-user-email-addresses-0889208.md "You can update the user email addresses used for logon.") +[Update User Email Addresses](update-user-email-addresses-0889208.md "You can update the user email addresses used for logon.") diff --git a/docs/Administering/Managing-Users-and-Roles/enabling-a-custom-saml-identity-provider-9b26536.md b/docs/Administering/Managing-Users-and-Roles/enabling-a-custom-saml-identity-provider-9b26536.md index 289feb4..ff66c10 100644 --- a/docs/Administering/Managing-Users-and-Roles/enabling-a-custom-saml-identity-provider-9b26536.md +++ b/docs/Administering/Managing-Users-and-Roles/enabling-a-custom-saml-identity-provider-9b26536.md @@ -303,7 +303,7 @@ SAP Datasphere can be hosted on non-SAP data centers. When dynamic user creation is enabled, new users will be automatically created using the default role and will be able to use SAML SSO to log onto SAP Datasphere. After users are created, you can set roles using SAML attributes. > ### Note: - > Automatic user deletion is not supported. If a user in SAP Datasphere is removed from your SAML IdP, you must go to *Security* \> *Users* and manually delete users. For more information, see [Deleting Users](deleting-users-3ceb94c.md). + > Automatic user deletion is not supported. If a user in SAP Datasphere is removed from your SAML IdP, you must go to *Security* \> *Users* and manually delete users. For more information, see [Delete Users](delete-users-3ceb94c.md). > > If this option is enabled, dynamic user creation still occurs even when SAML user attributes have not been set for all IdP users. To prevent a user from being automatically created, your SAML IdP must deny the user access to SAP Datasphere. diff --git a/docs/Administering/Managing-Users-and-Roles/exporting-users-e227d3c.md b/docs/Administering/Managing-Users-and-Roles/export-users-e227d3c.md similarity index 99% rename from docs/Administering/Managing-Users-and-Roles/exporting-users-e227d3c.md rename to docs/Administering/Managing-Users-and-Roles/export-users-e227d3c.md index cc76d1c..87103f2 100644 --- a/docs/Administering/Managing-Users-and-Roles/exporting-users-e227d3c.md +++ b/docs/Administering/Managing-Users-and-Roles/export-users-e227d3c.md @@ -2,7 +2,7 @@ -# Exporting Users +# Export Users If you want to synchronize SAP Datasphere user data with other systems, you can export the data to a CSV file. diff --git a/docs/Administering/Managing-Users-and-Roles/importing-or-modifying-users-from-a-file-b2698da.md b/docs/Administering/Managing-Users-and-Roles/import-or-modify-users-from-a-file-b2698da.md similarity index 96% rename from docs/Administering/Managing-Users-and-Roles/importing-or-modifying-users-from-a-file-b2698da.md rename to docs/Administering/Managing-Users-and-Roles/import-or-modify-users-from-a-file-b2698da.md index c1454cc..3ca3031 100644 --- a/docs/Administering/Managing-Users-and-Roles/importing-or-modifying-users-from-a-file-b2698da.md +++ b/docs/Administering/Managing-Users-and-Roles/import-or-modify-users-from-a-file-b2698da.md @@ -2,9 +2,9 @@ -# Importing or Modifying Users from a File +# Import or Modify Users from a File -You can create new users or batch-update existing users by importing user data that you have saved in a CSV file. +You can create users or batch-update existing users by importing user data that you have saved in a CSV file. @@ -14,7 +14,7 @@ The user data you want to import must be stored in a CSV file. At minimum, your If you want to assign new users different roles, include a `Roles` column in the CSV file. The role IDs used for role assignment are outlined in [Standard Roles Delivered with SAP Datasphere](standard-roles-delivered-with-sap-datasphere-a50a51d.md). -For existing users that you want to modify, you can create the CSV file by first exporting a CSV file from SAP Datasphere. For more information, see [Exporting Users](exporting-users-e227d3c.md). +For existing users that you want to modify, you can create the CSV file by first exporting a CSV file from SAP Datasphere. For more information, see [Export Users](export-users-e227d3c.md). > ### Note: > The first name, last name, and display name are linked to the identity provider, and can't be changed in the User list page, or when importing a CSV file. \(In the User list page, those columns are grayed out.\) diff --git a/docs/Administering/Managing-Users-and-Roles/managing-sap-datasphere-users-4fb82cb.md b/docs/Administering/Managing-Users-and-Roles/managing-sap-datasphere-users-4fb82cb.md index 8023463..b499a03 100644 --- a/docs/Administering/Managing-Users-and-Roles/managing-sap-datasphere-users-4fb82cb.md +++ b/docs/Administering/Managing-Users-and-Roles/managing-sap-datasphere-users-4fb82cb.md @@ -34,7 +34,7 @@ Create individual users in the *Users* list -[Creating a New User](creating-a-new-user-58d4b24.md) +[Create a User](create-a-user-58d4b24.md) @@ -46,7 +46,7 @@ Import multiple users from a CSV file -[Importing or Modifying Users from a File](importing-or-modifying-users-from-a-file-b2698da.md) +[Import or Modify Users from a File](import-or-modify-users-from-a-file-b2698da.md) @@ -80,7 +80,7 @@ Export user data to a CSV file, to synchronize with other systems -[Exporting Users](exporting-users-e227d3c.md) +[Export Users](export-users-e227d3c.md) @@ -92,7 +92,7 @@ Update the email address a user logs on with -[Updating User Email Addresses](updating-user-email-addresses-0889208.md) +[Update User Email Addresses](update-user-email-addresses-0889208.md) @@ -104,7 +104,7 @@ Delete users -[Deleting Users](deleting-users-3ceb94c.md) +[Delete Users](delete-users-3ceb94c.md) diff --git a/docs/Administering/Managing-Users-and-Roles/roles-and-privileges-by-app-and-feature-2d8b7d0.md b/docs/Administering/Managing-Users-and-Roles/roles-and-privileges-by-app-and-feature-2d8b7d0.md index 4618ab7..d0c19ef 100644 --- a/docs/Administering/Managing-Users-and-Roles/roles-and-privileges-by-app-and-feature-2d8b7d0.md +++ b/docs/Administering/Managing-Users-and-Roles/roles-and-privileges-by-app-and-feature-2d8b7d0.md @@ -925,7 +925,7 @@ DW Space Administrator Modify *Auditing* -See [Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who did what and when in the database.") :arrow_upper_right: +See [Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who performed which action at which point in time.") :arrow_upper_right: diff --git a/docs/Administering/Managing-Users-and-Roles/standard-roles-delivered-with-sap-datasphere-a50a51d.md b/docs/Administering/Managing-Users-and-Roles/standard-roles-delivered-with-sap-datasphere-a50a51d.md index 8275e85..71a17f4 100644 --- a/docs/Administering/Managing-Users-and-Roles/standard-roles-delivered-with-sap-datasphere-a50a51d.md +++ b/docs/Administering/Managing-Users-and-Roles/standard-roles-delivered-with-sap-datasphere-a50a51d.md @@ -8,7 +8,7 @@ SAP Datasphere is delivered with several standard roles. A standard role include -A DW Administrator can use standard roles as templates for creating custom roles with a different set of privileges \(see [Create a Custom Role](create-a-custom-role-862b88e.md)\). You can also use the standard roles that include scoped privileges as templates for creating scoped roles \(see [Create a Scoped Role to Assign Privileges to Users in Spaces](create-a-scoped-role-to-assign-privileges-to-users-in-spaces-b5c4e0b.md)\). You can assign the standard roles that contain global privileges \(DW Administrator, Catalog Administrator and Catalog User\) directly to users. +A DW Administrator can use standard roles as templates for creating custom roles with a different set of privileges \(see [Create a Custom Role](create-a-custom-role-862b88e.md)\). You can also use the standard roles that include scoped privileges as templates for creating scoped roles \(see [Create a Scoped Role to Assign Privileges to Users in Spaces](create-a-scoped-role-to-assign-privileges-to-users-in-spaces-b5c4e0b.md)\). You can assign the standard roles that contain global privileges \(such as DW Administrator, Catalog Administrator and Catalog User\) directly to users. > ### Note: > You cannot delete nor edit standard roles. diff --git a/docs/Administering/Managing-Users-and-Roles/updating-user-email-addresses-0889208.md b/docs/Administering/Managing-Users-and-Roles/update-user-email-addresses-0889208.md similarity index 70% rename from docs/Administering/Managing-Users-and-Roles/updating-user-email-addresses-0889208.md rename to docs/Administering/Managing-Users-and-Roles/update-user-email-addresses-0889208.md index 9703db6..2c2a943 100644 --- a/docs/Administering/Managing-Users-and-Roles/updating-user-email-addresses-0889208.md +++ b/docs/Administering/Managing-Users-and-Roles/update-user-email-addresses-0889208.md @@ -1,6 +1,6 @@ -# Updating User Email Addresses +# Update User Email Addresses You can update the user email addresses used for logon. @@ -17,7 +17,7 @@ As long as a user has not logged on to the system with the new email address, th **Related Information** -[Creating a New User](creating-a-new-user-58d4b24.md "You can create individual users in SAP Datasphere.") +[Create a User](create-a-user-58d4b24.md "You can create individual users in SAP Datasphere.") -[Importing or Modifying Users from a File](importing-or-modifying-users-from-a-file-b2698da.md "You can create new users or batch-update existing users by importing user data that you have saved in a CSV file.") +[Import or Modify Users from a File](import-or-modify-users-from-a-file-b2698da.md "You can create users or batch-update existing users by importing user data that you have saved in a CSV file.") diff --git a/docs/Administering/Monitoring-SAP-Datasphere/analyze-monitoring-data-in-a-space-9cd0691.md b/docs/Administering/Monitoring-SAP-Datasphere/analyze-monitoring-data-in-a-space-9cd0691.md index 8339e7b..80c9bb5 100644 --- a/docs/Administering/Monitoring-SAP-Datasphere/analyze-monitoring-data-in-a-space-9cd0691.md +++ b/docs/Administering/Monitoring-SAP-Datasphere/analyze-monitoring-data-in-a-space-9cd0691.md @@ -82,7 +82,7 @@ In the area *Expensive Statement Tracing*, you can enable expensive statement tr - The information on statements that exceed the specified thresholds are included in dedicated views that you can access in the selected monitoring space. -- All the database statements that exceed the thresholds specified for memory consumption \(*Threshold Memory*\) and runtime \(*Threshold Duration*\) are displayed in the widgets and tables of the *System Monitor*. If expensive statement tracing is not enabled, then statement information and errors are not traced and you cannot see them in the *System Monitor* \(see [Monitoring SAP Datasphere](monitoring-sap-datasphere-28910cd.md)\). +- All the database statements that exceed the thresholds specified for memory consumption \(*Threshold Memory*\) and runtime \(*Threshold Duration*\) are displayed in the cards and tables of the *System Monitor*. If expensive statement tracing is not enabled, then statement information and errors are not traced and you cannot see them in the *System Monitor* \(see [Monitoring SAP Datasphere](monitoring-sap-datasphere-28910cd.md)\). - By default, 30 000 records maximum are stored in the monitoring tables. You can change this number, which will impact the views dedicated to monitoring and information related to statements in the *System Monitor*. For example, if about 5 days are traced in the expensive statement tables and you don’t want to change the thresholds, you can double the number of records in *In-Memory Tracing Records* so that about 10 days are traced. Be aware that increasing this number will also increase the used storage. @@ -114,7 +114,7 @@ To analyze individual elastic compute node query, you can enable the tracing of ## Procedure -1. Go to \(Configuration\) → *Monitoring*. +1. In the side navigation area, click \(*System*\) ** \> ** :wrench: \(*Configuration*\)** \> *Monitoring*. 2. Select a space from the drop-down list and click *Confirm Selected Space*. 3. If you've created the ** space and you want to enable it, click *Enable access to SAP Monitoring Content Space*. If there isn't any space named ** in your tenant, this is not available for selection. 4. To trace expensive statements, select *Enable Expensive Statement Tracing*, specify the following parameters to configure and filter the trace details, then save your changes. diff --git a/docs/Administering/Monitoring-SAP-Datasphere/create-a-database-analysis-user-to-debug-database-issues-c28145b.md b/docs/Administering/Monitoring-SAP-Datasphere/create-a-database-analysis-user-to-debug-database-issues-c28145b.md index 83b8c31..9a6e1c1 100644 --- a/docs/Administering/Monitoring-SAP-Datasphere/create-a-database-analysis-user-to-debug-database-issues-c28145b.md +++ b/docs/Administering/Monitoring-SAP-Datasphere/create-a-database-analysis-user-to-debug-database-issues-c28145b.md @@ -95,7 +95,7 @@ A user with the *DW Administrator* role can create a database analysis user. > ### Note: - > All actions of the database analysis user are logged in the `ANALYSIS_AUDIT_LOG` view, which is stored in the space that has been assigned to store audit logs \(see [Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who did what and when in the database.") :arrow_upper_right:\). + > All actions of the database analysis user are logged in the `ANALYSIS_AUDIT_LOG` view, which is stored in the space that has been assigned to store audit logs \(see [Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who performed which action at which point in time.") :arrow_upper_right:\). > > The audit logs entries are kept for 180 days, after which they are deleted. diff --git a/docs/Administering/Monitoring-SAP-Datasphere/monitor-database-operations-with-audit-logs-110404a.md b/docs/Administering/Monitoring-SAP-Datasphere/monitor-database-operations-with-audit-logs-110404a.md index 78d3b57..8880ce9 100644 --- a/docs/Administering/Monitoring-SAP-Datasphere/monitor-database-operations-with-audit-logs-110404a.md +++ b/docs/Administering/Monitoring-SAP-Datasphere/monitor-database-operations-with-audit-logs-110404a.md @@ -4,7 +4,7 @@ Monitor the read and change actions \(policies\) performed in the database with audit logs, and see who did what and when. -If Space Administrators have enabled audit logs to be created for their space \(see [Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who did what and when in the database.") :arrow_upper_right:\), you can get an overview of these audit logs. You can do analytics on audit logs by assigning the audit views to a dedicated space and then work with them in a view in the Data Builder. +If Space Administrators have enabled audit logs to be created for their space \(see [Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who performed which action at which point in time.") :arrow_upper_right:\), you can get an overview of these audit logs. You can do analytics on audit logs by assigning the audit views to a dedicated space and then work with them in a view in the Data Builder. > ### Note: > Audit logs can consume a large quantity of GB of disk in your database, especially when combined with long retention periods \(which are defined at the space level\). You can delete audit logs when needed, which will free up disk space. For more information, see [Delete Audit Logs](delete-audit-logs-589fa42.md). diff --git a/docs/Administering/Monitoring-SAP-Datasphere/monitoring-sap-datasphere-28910cd.md b/docs/Administering/Monitoring-SAP-Datasphere/monitoring-sap-datasphere-28910cd.md index fc6989f..9047e6c 100644 --- a/docs/Administering/Monitoring-SAP-Datasphere/monitoring-sap-datasphere-28910cd.md +++ b/docs/Administering/Monitoring-SAP-Datasphere/monitoring-sap-datasphere-28910cd.md @@ -29,7 +29,7 @@ For example, you can see all the errors \(such as failed tasks and out-of-memory You can monitor out-of-memory errors and other information that are related to SAP HANA database SQL statements, depending on what you've specified in \(Configuration\) → *Monitoring*: -- If *Enable Expensive Statement Tracing* is not selected, then in *System Monitor* \> *Dashboard*, you cannot see the widgets about out-of-memory errors and about other information related to statements. For example, you cannot see the widgets: *Out-of-Memory Errors*, *Top 5 Statements by Processing Memory Consumption*. +- If *Enable Expensive Statement Tracing* is not selected, then in *System Monitor* \> *Dashboard*, you cannot see the cards about out-of-memory errors and about other information related to statements. For example, you cannot see the cards: *Out-of-Memory Errors*, *Top 5 Statements by Processing Memory Consumption*. - If *Enable Expensive Statement Tracing* is not selected and none of the threshold options is selected, then in the tables of *System Monitor* \> *Logs*, you cannot see any information about out-of-memory errors and about other information related to statements. For example, no information is displayed in the columns *Peak Memory \(MiB\)* and *Peak CPU \(ms\)*. @@ -55,7 +55,7 @@ For more information on enabling and configuring expensive statement tracing, se ## Dashboard -The out-of-memory widgets and top-memory consumption widgets help you to identify issues. You can see detailed information about the issues by clicking *View Logs*, which takes you to the *Logs* tab. For example, you can find out if tasks have to be scheduled at another time so that high-memory consuming tasks do not run at the same time. If single tasks consume too much memory, some additional views may need to be persisted or the view partitioning may need to be used to lower the memory consumption. +The out-of-memory cards and top-memory consumption cards help you to identify issues. You can see detailed information about the issues by clicking *View Logs*, which takes you to the *Logs* tab. For example, you can find out if tasks have to be scheduled at another time so that high-memory consuming tasks do not run at the same time. If single tasks consume too much memory, some additional views may need to be persisted or the view partitioning may need to be used to lower the memory consumption. The following information is available in the *Dashboard* tab: @@ -79,14 +79,14 @@ The following information is available in the *Dashboard* tab: - *Disk Assigned to Spaces for Storage* - Shows the total amount of disk storage assigned to all spaces. - *Memory Assigned to Spaces for Storage* - Shows the total amount of memory storage assigned to all spaces. -For each of the key indicator widgets listed below, you can see detailed information by clicking the link *View Logs*, which takes you to the *Logs* tab. +For each of the key indicator cards listed below, you can see detailed information by clicking the link *View Logs*, which takes you to the *Logs* tab.
-Widget +Card @@ -345,7 +345,7 @@ Shows the number of statements that have been queued in the last 7 days because -Shows the 5 spaces with the highest number of rejected statements in the last 24 hours. +Shows the 5 spaces with the highest number of rejected statements in the last 7 days. > ### Note: > A space that has been deleted is prefixed with an asterisk character. @@ -364,7 +364,7 @@ Shows the 5 spaces with the highest number of rejected statements in the last 24 -Shows the 5 spaces with the highest number of queued statements in the last 24 hours. +Shows the 5 spaces with the highest number of queued statements in the last 7 days. > ### Note: > A space that has been deleted is prefixed with an asterisk character. @@ -376,7 +376,7 @@ Shows the 5 spaces with the highest number of queued statements in the last 24 h
> ### Note: -> - To investigate why statements are being queued or rejected, you can click *Open SAP HANA Cockpit* in the widgets dedicated to admission contol. If you've created a database analysis user, you're connected to the SAP HANA Cockpit without entering your credentials \(see [Create a Database Analysis User to Debug Database Issues](create-a-database-analysis-user-to-debug-database-issues-c28145b.md). +> - To investigate why statements are being queued or rejected, you can click *Open SAP HANA Cockpit* in the cards dedicated to admission contol. If you've created a database analysis user, you're connected to the SAP HANA Cockpit without entering your credentials \(see [Create a Database Analysis User to Debug Database Issues](create-a-database-analysis-user-to-debug-database-issues-c28145b.md). > > - For more information about admission control thresholds, see [Set a Priority and Statement Limits for a Space](../Creating-Spaces-and-Allocating-Storage/set-a-priority-and-statement-limits-for-a-space-d66ac1e.md). diff --git a/docs/Administering/Preparing-Connectivity/configure-cloud-connector-f289920.md b/docs/Administering/Preparing-Connectivity/configure-cloud-connector-f289920.md index a776915..e3dabfc 100644 --- a/docs/Administering/Preparing-Connectivity/configure-cloud-connector-f289920.md +++ b/docs/Administering/Preparing-Connectivity/configure-cloud-connector-f289920.md @@ -269,6 +269,18 @@ For more information about the supported use cases depending on the connection t + *Oracle* \(data flows\) + + + + + TCP + + + + + + *Apache Kafka* on-premise only \(replication flows\) diff --git a/docs/Administering/Preparing-Connectivity/prepare-connectivity-to-oracle-9fca7c4.md b/docs/Administering/Preparing-Connectivity/prepare-connectivity-to-oracle-9fca7c4.md index cd31404..2a3dc74 100644 --- a/docs/Administering/Preparing-Connectivity/prepare-connectivity-to-oracle-9fca7c4.md +++ b/docs/Administering/Preparing-Connectivity/prepare-connectivity-to-oracle-9fca7c4.md @@ -29,7 +29,12 @@ Before you can use the connection for creating views and accessing data via remo Before you can use the connection for data flows, the following is required: -- To directly consume data in data flows, the Oracle database must be available on the public internet. +- An administrator has installed and configured Cloud Connector to connect to your on-premise source. + + For more information, see [Configure Cloud Connector](configure-cloud-connector-f289920.md). + + > ### Note: + > Cloud Connector is not required if your Oracle database is available on the public internet. - A DW administrator has uploaded the required ODBC driver file to SAP Datasphere. @@ -45,5 +50,5 @@ Before you can use the connection for data flows, the following is required: **Related Information** -[Oracle Connections](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/c73ae0601d364f47830d339b6e86b7e8.html "Use an Oracle connection to access data from an Oracle database (on-premise).") :arrow_upper_right: +[Oracle Connections](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/c73ae0601d364f47830d339b6e86b7e8.html "Use the connection to connect to and access data from an Oracle database (on-premise).") :arrow_upper_right: diff --git a/docs/Administering/Preparing-Connectivity/prepare-connectivity-to-sap-s-4hana-cloud-abb159e.md b/docs/Administering/Preparing-Connectivity/prepare-connectivity-to-sap-s-4hana-cloud-abb159e.md index 637c0ea..1bea56b 100644 --- a/docs/Administering/Preparing-Connectivity/prepare-connectivity-to-sap-s-4hana-cloud-abb159e.md +++ b/docs/Administering/Preparing-Connectivity/prepare-connectivity-to-sap-s-4hana-cloud-abb159e.md @@ -21,6 +21,8 @@ Before you can use the connection for creating views and accessing data via remo For more information, see [Integrating CDI](https://help.sap.com/viewer/0f69f8fb28ac4bf48d2b57b9637e81fa/latest/en-US/4a006b43551d4cb5aed6399c0ace6b98.html) in the *SAP S/4HANA Cloud* documentation. +If you want to use federated access to CDS view entities using the ABAP SQL service exposure from SAP S/4HANA Cloud, see [Using ABAP SQL Services for Accessing Data from SAP S/4HANA Cloud](using-abap-sql-services-for-accessing-data-from-sap-s-4hana-cloud-ef2b223.md). +
@@ -44,26 +46,12 @@ Before you can use the connection for data flows, the following is required: Before you can use the connection for replication flows, the following is required: -- If you want to replicate CDS views: - - A communication arrangement has been created for communication scenario `SAP_COM_0532` in the SAP S/4HANA Cloud system. +- A communication arrangement has been created for communication scenario `SAP_COM_0532` in the SAP S/4HANA Cloud system. For more information, see [Integrating CDS Views Using SAP Datasphere](https://help.sap.com/viewer/0f69f8fb28ac4bf48d2b57b9637e81fa/latest/en-US/f509eddda867452db9631dae1ae442a3.html) in the *SAP S/4HANA Cloud* documentation. -- If you want to replicate CDS view entities using the SQL service exposure from SAP S/4HANA Cloud: - - In SAP S/4HANA Cloud, your system administration has created the relevant communication arrangements: - - - Communication arrangements for exposing SQL services \(see [Exposing the SQL Service for Data Federation and Replication with Privileged Access](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/70b2fc2a9e37475b993d4e6fd6d3eb07.html) in the *SAP S/4HANA Cloud* documentation\) - - - A communication arrangement for communication scenario `SAP_COM_0532` \(see [Integrating SQL Services Using SAP Datasphere](https://help.sap.com/docs/SAP_S4HANA_CLOUD/0f69f8fb28ac4bf48d2b57b9637e81fa/a91d39c3ea494344897fca7bc9578b85.html) in the *SAP S/4HANA Cloud* documentation\) - - - > ### Note: - > Note that the same communication user must be added to all communication arrangements. - - For more information about using SQL services to replicate ABAP-managed data to SAP Datasphere, see [Data Consumption Using SAP Datasphere](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/ec312dd3e39f401b84681c53adc08ad8.html) in the *SAP S/4HANA Cloud* documentation. +If you want to replicate CDS view entities using the ABAP SQL service exposure from SAP S/4HANA Cloud, see [Using ABAP SQL Services for Accessing Data from SAP S/4HANA Cloud](using-abap-sql-services-for-accessing-data-from-sap-s-4hana-cloud-ef2b223.md). diff --git a/docs/Administering/Preparing-Connectivity/preparing-connectivity-for-connections-bffbd58.md b/docs/Administering/Preparing-Connectivity/preparing-connectivity-for-connections-bffbd58.md index 5dd1f16..f08e48a 100644 --- a/docs/Administering/Preparing-Connectivity/preparing-connectivity-for-connections-bffbd58.md +++ b/docs/Administering/Preparing-Connectivity/preparing-connectivity-for-connections-bffbd58.md @@ -855,7 +855,7 @@ no -[Oracle Connections](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/c73ae0601d364f47830d339b6e86b7e8.html "Use an Oracle connection to access data from an Oracle database (on-premise).") :arrow_upper_right: +[Oracle Connections](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/c73ae0601d364f47830d339b6e86b7e8.html "Use the connection to connect to and access data from an Oracle database (on-premise).") :arrow_upper_right: @@ -870,7 +870,7 @@ yes -no +yes \(for data flows\) diff --git a/docs/Administering/Preparing-Connectivity/upload-third-party-odbc-drivers-required-for-data-flows-b9b5579.md b/docs/Administering/Preparing-Connectivity/upload-third-party-odbc-drivers-required-for-data-flows-b9b5579.md index 1698fb1..2e242d9 100644 --- a/docs/Administering/Preparing-Connectivity/upload-third-party-odbc-drivers-required-for-data-flows-b9b5579.md +++ b/docs/Administering/Preparing-Connectivity/upload-third-party-odbc-drivers-required-for-data-flows-b9b5579.md @@ -86,7 +86,7 @@ ee79a8d41760a90b6fa2e1a074e33b0518e3393afd305f0bee843b5393e10df0 -[Oracle Connections](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/c73ae0601d364f47830d339b6e86b7e8.html "Use an Oracle connection to access data from an Oracle database (on-premise).") :arrow_upper_right: +[Oracle Connections](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/c73ae0601d364f47830d339b6e86b7e8.html "Use the connection to connect to and access data from an Oracle database (on-premise).") :arrow_upper_right: @@ -121,7 +121,7 @@ instantclient-basiclite-linux.x64-19.17.0.0.0dbru.zip > - Additional files are required if SSL is used: > > - oraclepki.jar \(SHA256 fingerprint: e408e7ae67650917dbce3ad263829bdc6c791d50d4db2fd59aeeb5503175499b\) -> - package zip file. The package applies to allosdt\_cert.jar \(SHA256 fingerprint: 6b152d4332bd39f258a88e58b9215a926048d740e148971fe1628b09060176a8\) +> - osdt\_cert.jar \(SHA256 fingerprint: 6b152d4332bd39f258a88e58b9215a926048d740e148971fe1628b09060176a8\) > - osdt\_core.jar \(SHA256 fingerprint: c25e30184bb94c6da1227c8256f0e1336acb97b29229edb4aacf27167b96075e\) diff --git a/docs/Administering/Preparing-Connectivity/using-abap-sql-services-for-accessing-data-from-sap-s-4hana-cloud-ef2b223.md b/docs/Administering/Preparing-Connectivity/using-abap-sql-services-for-accessing-data-from-sap-s-4hana-cloud-ef2b223.md new file mode 100644 index 0000000..e1f5801 --- /dev/null +++ b/docs/Administering/Preparing-Connectivity/using-abap-sql-services-for-accessing-data-from-sap-s-4hana-cloud-ef2b223.md @@ -0,0 +1,94 @@ + + +# Using ABAP SQL Services for Accessing Data from SAP S/4HANA Cloud + +The ABAP SQL service provides SQL-level access to published CDS view entities for SAP Datasphere. You can use the service to replicate data with replication flows or to federate data with remote tables. + +> ### Note: +> This feature is only available in an SAP S/4HANA Cloud 3-system landscape. For information about system landscapes, see: [System Landscapes in SAP S/4HANA Cloud](https://help.sap.com/docs/SAP_S4HANA_CLOUD/a630d57fc5004c6383e7a81efee7a8bb/aa60b129af7b4ce8ae052618c8315d29.html) in the *SAP S/4HANA Cloud* documentation. + +For both consumption scenarios using the SQL service, data federation and data replication, privileged data access needs to be enabled for communication users in SAP S/4HANA Cloud. + +For more information about the consumption scenarios and privileged access, see [Data Integration Patterns](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/96368bd086ff4f79933b078a6cf7feaa.html) in the *SAP S/4HANA Cloud* documentation. + + + + + +## Data Federation With Remote Tables + +In SAP S/4HANA Cloud, a business user and administrator must perform the following steps to prepare data federation with remote tables: + +1. There are some prerequisites and constraints that must be considered before using the SQL service. + + For more information, see [Prerequisites and Constraints](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/5e8488149308423b9f8b2f2c8ef9c761.html) in the *SAP S/4HANA Cloud* documentation. Note that the information about the ODBC driver is not relevant for SAP Datasphere as a consumer of an exposed SQL service. + +2. To expose CDS view entities using the SQL service, an SAP S/4HANA Cloud business user has created a service definition and a corresponding service binding of type SQL1 in the ABAP Development Tools. The service definition lists the set of CDS view entities that shall be exposed, and a service binding of type SQL for that service definition enables their exposure via the ABAP SQL Service. + + In the *Enabled Operations* area of the service binding, the business user must select access type *SELECT* to enable federated access. + + For more information, see [Creating a Service Definition and an SQL-Typed Service Binding](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/c1cf6c9796ad4fecb893672fd91e660d.html) in the *SAP S/4HANA Cloud* documentation. + +3. To expose the SQL service to get privileged access to the CDS view entities with a communication user, a communication arrangement is required. This involves the following steps: + + 1. An SAP S/4HANA Cloud business user has created a custom communication scenario in the ABAP Development Tools. + + When filling out the authorizations for authorization object `S_SQL_VIEW` in the communication scenario, note the following: + + - On the *Sources* tab of the Data Builder view editors in SAP Datasphere, the service binding name from the *SQL\_SCHEMA* authorization field is visible as \(virtual\) schema. + + - In the SQL\_VIEWOP authorization field, select the option *SELECT* to grant federated access. + + + 2. An administrator has created a communication system and user in the SAP Fiori launchpad of the ABAP environment. + + 3. An administrator has created a communication arrangement for exposing the SQL service in the SAP Fiori launchpad of the ABAP environment. + + + For more information about the above steps, see [Exposing the SQL Service for Data Federation and Replication with Privileged Access](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/70b2fc2a9e37475b993d4e6fd6d3eb07.html) in the *SAP S/4HANA Cloud* documentation. + + + + + + +## Data Replication With Replication Flows + +In SAP S/4HANA Cloud, a business user and administrator must perform the following steps to prepare data replication with replication flows: + +1. There are some prerequisites and constraints that must be considered before using the SQL service. + + For more information, see [Prerequisites and Constraints](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/5e8488149308423b9f8b2f2c8ef9c761.html) in the *SAP S/4HANA Cloud* documentation. Note that the information about the ODBC driver is not relevant for SAP Datasphere as a consumer of an exposed SQL service. + +2. To expose CDS view entities using the SQL service, an SAP S/4HANA Cloud business user has created a service definition and a corresponding service binding of type SQL1 in the ABAP Development Tools. The service definition lists the set of CDS view entities that shall be exposed, and a service binding of type SQL for that service definition enables their exposure via the ABAP SQL Service. + + In the *Enabled Operations* area of the service binding, the business user must select access type *REPLICATE* to enable data replication. + + For more information, see [Creating a Service Definition and an SQL-Typed Service Binding](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/c1cf6c9796ad4fecb893672fd91e660d.html) in the *SAP S/4HANA Cloud* documentation. + +3. To expose the SQL service to get privileged access to the CDS view entities with a communication user, a communication arrangement is required. This involves the following steps: + + 1. An SAP S/4HANA Cloud business user has created a custom communication scenario in the ABAP Development Tools. + + When filling out the authorizations for authorization object `S_SQL_VIEW` in the communication scenario, note the following: + + - In the SQL\_VIEWOP authorization field, select the option *REPLICATE* to allow replication on the specified views. + + + 2. An administrator has created a communication system and user in the SAP Fiori launchpad of the ABAP environment. + + 3. An administrator has created a communication arrangement for exposing the SQL service in the SAP Fiori launchpad of the ABAP environment. + + + For more information about the above steps, see [Exposing the SQL Service for Data Federation and Replication with Privileged Access](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/70b2fc2a9e37475b993d4e6fd6d3eb07.html) in the *SAP S/4HANA Cloud* documentation. + +4. An administrator has created a communication arrangement for communication scenario SAP\_COM\_0532 in the SAP Fiori launchpad of the ABAP environment. + + For more information, see [Creating a Communication Arrangement to Enable Replication Flows in SAP Datasphere](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/96368bd086ff4f79933b078a6cf7feaa.html) in the *SAP S/4HANA Cloud* documentation. + + +> ### Note: +> Note that the same communication user must be added to all communication arrangements. + +For more information about using SQL services to replicate ABAP-managed data to SAP Datasphere, see [Data Consumption Using SAP Datasphere](https://help.sap.com/docs/SAP_S4HANA_CLOUD/6aa39f1ac05441e5a23f484f31e477e7/ec312dd3e39f401b84681c53adc08ad8.html) in the *SAP S/4HANA Cloud* documentation. + diff --git a/docs/Administering/administration-apps-and-tools-c6dd052.md b/docs/Administering/administration-apps-and-tools-c6dd052.md index 4b41f3a..1d95450 100644 --- a/docs/Administering/administration-apps-and-tools-c6dd052.md +++ b/docs/Administering/administration-apps-and-tools-c6dd052.md @@ -233,7 +233,7 @@ More Information -[Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who did what and when in the database.") :arrow_upper_right: +[Enable Audit Logging](https://help.sap.com/viewer/9f36ca35bc6145e4acdef6b4d852d560/DEV_CURRENT/en-US/266553976e1c4db9aaa28a75e2308b77.html "You can enable audit logs for your space so that read and change actions (policies) are recorded. Administrators can then analyze who performed which action at which point in time.") :arrow_upper_right: diff --git a/docs/Administering/index.md b/docs/Administering/index.md index 022e7b6..dab7862 100644 --- a/docs/Administering/index.md +++ b/docs/Administering/index.md @@ -28,11 +28,11 @@ - [Updating the SAML IdP Signing Certificate](Managing-Users-and-Roles/updating-the-saml-idp-signing-certificate-ba968db.md) - [Identity Provider Administration](Managing-Users-and-Roles/identity-provider-administration-df15ed8.md) - [Managing SAP Datasphere Users](Managing-Users-and-Roles/managing-sap-datasphere-users-4fb82cb.md) - - [Creating a New User](Managing-Users-and-Roles/creating-a-new-user-58d4b24.md) - - [Importing or Modifying Users from a File](Managing-Users-and-Roles/importing-or-modifying-users-from-a-file-b2698da.md) - - [Exporting Users](Managing-Users-and-Roles/exporting-users-e227d3c.md) - - [Updating User Email Addresses](Managing-Users-and-Roles/updating-user-email-addresses-0889208.md) - - [Deleting Users](Managing-Users-and-Roles/deleting-users-3ceb94c.md) + - [Create a User](Managing-Users-and-Roles/create-a-user-58d4b24.md) + - [Import or Modify Users from a File](Managing-Users-and-Roles/import-or-modify-users-from-a-file-b2698da.md) + - [Export Users](Managing-Users-and-Roles/export-users-e227d3c.md) + - [Update User Email Addresses](Managing-Users-and-Roles/update-user-email-addresses-0889208.md) + - [Delete Users](Managing-Users-and-Roles/delete-users-3ceb94c.md) - [Set a Password Policy for Database Users](Managing-Users-and-Roles/set-a-password-policy-for-database-users-14aedf6.md) - [Managing Roles and Privileges](Managing-Users-and-Roles/managing-roles-and-privileges-3740dac.md) - [Standard Roles Delivered with SAP Datasphere](Managing-Users-and-Roles/standard-roles-delivered-with-sap-datasphere-a50a51d.md) @@ -92,6 +92,7 @@ - [Prepare Connectivity to SAP Marketing Cloud](Preparing-Connectivity/prepare-connectivity-to-sap-marketing-cloud-f5e0c06.md) - [Prepare Connectivity to SAP SuccessFactors](Preparing-Connectivity/prepare-connectivity-to-sap-successfactors-c9b1915.md) - [Prepare Connectivity to SAP S/4HANA Cloud](Preparing-Connectivity/prepare-connectivity-to-sap-s-4hana-cloud-abb159e.md) + - [Using ABAP SQL Services for Accessing Data from SAP S/4HANA Cloud](Preparing-Connectivity/using-abap-sql-services-for-accessing-data-from-sap-s-4hana-cloud-ef2b223.md) - [Prepare Connectivity to SAP S/4HANA On-Premise](Preparing-Connectivity/prepare-connectivity-to-sap-s-4hana-on-premise-8de01dd.md) - [Create SAP S/4HANA Live Data Connection of Type Tunnel](Preparing-Connectivity/create-sap-s-4hana-live-data-connection-of-type-tunnel-095dbdf.md) - [Managing and Monitoring Connectivity for Data Integration](managing-and-monitoring-connectivity-for-data-integration-c5b167b.md) diff --git a/docs/Administering/receive-notifications-about-data-provisioning-agent-status-changes-85790bb.md b/docs/Administering/receive-notifications-about-data-provisioning-agent-status-changes-85790bb.md index 383cc9e..f3c3b6b 100644 --- a/docs/Administering/receive-notifications-about-data-provisioning-agent-status-changes-85790bb.md +++ b/docs/Administering/receive-notifications-about-data-provisioning-agent-status-changes-85790bb.md @@ -30,19 +30,17 @@ A recurring task will check for any status changes according to the configured f 1. In the side navigation area, click \(*System*\) ** \> ** :wrench: \(*Configuration*\) ** \> *Data Integration*. -2. +2. Go to the *On-Premise Agents* section and click \(menu\) ** \> ** *Configure Sending Notifications*. -3. Go to the *On-Premise Agents* section and click \(menu\) ** \> ** *Configure Sending Notifications*. +3. If you haven't authorized SAP Datasphere yet to run your scheduled tasks for you, you will see a message at the top of the *Configure Sending Notifications* dialog asking for your consent. Give your consent. -4. If you haven't authorized SAP Datasphere yet to run your scheduled tasks for you, you will see a message at the top of the *Configure Sending Notifications* dialog asking for your consent. Give your consent. - -5. Switch on the *Send Notifications* toggle. +4. Switch on the *Send Notifications* toggle. An additional field *Owner* appears that shows that you have been automatically assigned as the owner of the task. -6. Select the frequency in which the status of the Data Provisioning Agent should be checked. +5. Select the frequency in which the status of the Data Provisioning Agent should be checked. -7. Save your configuration. +6. Save your configuration. This will start the first status check. After the first check, the status check will be performed according to the defined frequency. diff --git a/docs/Administering/request-help-from-sap-support-831a977.md b/docs/Administering/request-help-from-sap-support-831a977.md index ceca74b..1192434 100644 --- a/docs/Administering/request-help-from-sap-support-831a977.md +++ b/docs/Administering/request-help-from-sap-support-831a977.md @@ -10,43 +10,52 @@ You can create an SAP support incident on the [SAP Support Portal](https://suppo An administrator can make sure that a support user is created in your tenant. Two options are available: -- An administrator creates the support user. +- An administrator generally allows SAP Product Support to create support users based on incidents. - Before creating an incident with SAP, proceed as follows: + Proceed as follows: - 1. In the shell bar, click \(*Support*\). + 1. In the side navigation area, click \(*System*\) ** \> ** \(*Administration*\) ** \> *System Configuration*. - 2. In the *Support* dialog, click *Create Support User* and then choose *OK* to confirm the support user creation. + > ### Note: + > If your tenant was provisioned prior to version 2021.03, click \(Product Switch\) ** \> ** *Analytics*** \> ** *System* \> *Administration* \> *System Configuration*. - An email is automatically sent to SAP Support to notify them of the newly created support user, and it is listed with your other users at *Security* \> *Users*. + 2. Choose *Edit*. - The support user has minimum privileges and does not consume a user license. + 3. Set the *Allow SAP support user creation* setting to *ON*. - You can assign an appropriate role to the support user \(DW Administrator role\) and add it to the required space. + 4. Click *Save*. - 3. Delete the support user when your issue is resolved. + In case of an incident, the assigned support engineer from SAP Product Support can request and generate a personalized support user for the affected tenant. This user is enabled for multi-factor authentication. + Support engineers can request the support user with one of the following roles: -- An administrator generally allows SAP Product Support to create support users based on incidents. + - the global extended role *DW Support User* along with the scoped role *DW Scoped Support User* - Proceed as follows: + *DW Support User* gives support users read-only access privileges to all functionalities of SAP Datasphere, enabling them to analyze the incident. - 1. In the side navigation area, click \(*System*\) ** \> ** \(*Administration*\) ** \> *System Configuration*. + When support engineers request the *DW Scoped Support User* role, they can specify the spaces that need to be added as scopes to this role. This gives the support user read-only access to these spaces. - > ### Note: - > If your tenant was provisioned prior to version 2021.03, click \(Product Switch\) ** \> ** *Analytics*** \> ** *System* \> *Administration* \> *System Configuration*. + - the global *DW Administrator* role, if the customer confirms this in the incident - 2. Choose *Edit*. - 3. Set the *Allow SAP support user creation* setting to *ON*. + The support user does not consume a user license, and it will be automatically deleted after two days or after the incident has been closed. - 4. Click *Save*. - In case of an incident, a support engineer from SAP Product Support can now request and generate a personalized support user for the affected tenant. This user is enabled for multi-factor authentication. +- An administrator creates the support user. + + Before creating an incident with SAP, proceed as follows: + + 1. In the shell bar, click \(*Support*\). - The support user either has read-only privileges \(DW Viewer role\) or, if requested by the support engineer and confirmed by the customer in the incident, the support user can have the DW Administrator role. The user does not consume a user license. + 2. In the *Support* dialog, click *Create Support User* and then choose *OK* to confirm the support user creation. - The support user will be automatically deleted after two days or after the incident has been closed. + An email is automatically sent to SAP Support to notify them of the newly created support user, and it is listed with your other users at *Security* \> *Users*. + + The support user has minimum privileges and does not consume a user license. + + You can assign an appropriate role to the support user \(DW Administrator role\) and add it to the required space. + + 3. Delete the support user when your issue is resolved. diff --git a/docs/Integrating-data-and-managing-spaces/Data-Integration-Monitor/creating-statistics-for-your-remote-tables-e4120bb.md b/docs/Integrating-data-and-managing-spaces/Data-Integration-Monitor/creating-statistics-for-your-remote-tables-e4120bb.md index 392131a..28111d0 100644 --- a/docs/Integrating-data-and-managing-spaces/Data-Integration-Monitor/creating-statistics-for-your-remote-tables-e4120bb.md +++ b/docs/Integrating-data-and-managing-spaces/Data-Integration-Monitor/creating-statistics-for-your-remote-tables-e4120bb.md @@ -95,7 +95,7 @@ Creates a data statistics object that helps the query optimizer estimate the dat Statistics are created for each column of the table and can only be created if the data access is "Remote". Only one data statistics object at a time is allowed per remote table. > ### Note: -> Creating statistics can cause a significant workload on the source system. RECORD COUNT is the simplest statistics object, and complexity increases for SIMPLE and HISTOGRAM, which causes heavier workload. +> Creating statistics can cause a significant workload on the source system, especially if the source object is a view, like for example a CDS View in ABAP or an SQL View in a Database. RECORD COUNT is the simplest statistics object, and complexity increases for SIMPLE and HISTOGRAM, which causes heavier workload. You can update data statistics objects: when you choose the menu *Create Statistics* on a remote table for which data statistics are already available, the windows *Create Statistics* opens, and the existing type of statistics line is marked as active. You can then choose to create another type of statistics or update the existing ones. diff --git a/docs/Integrating-data-and-managing-spaces/Data-Integration-Monitor/monitoring-local-tables-3740461.md b/docs/Integrating-data-and-managing-spaces/Data-Integration-Monitor/monitoring-local-tables-3740461.md index 34157f1..a97ae05 100644 --- a/docs/Integrating-data-and-managing-spaces/Data-Integration-Monitor/monitoring-local-tables-3740461.md +++ b/docs/Integrating-data-and-managing-spaces/Data-Integration-Monitor/monitoring-local-tables-3740461.md @@ -112,6 +112,11 @@ Displays the memory currently used by the local table in MiB. Displays the timestamp when the local table was last updated. +> ### Restriction: +> If the local table is updated through a replication flow, the information is not available as not supported for this flow type. + + + @@ -127,6 +132,9 @@ Indicates how the table was last updated. For example, name of the flow if it wa > ### Note: > If you are authorized to navigate to the relevant object, a clickable link is provided and you can navigate to the object details. If you are not authorized to navigate to the object, you’ll get the value unauthorized. +> ### Restriction: +> If the local table is updated through a replication flow, the information is not available as not supported for this flow type. + diff --git a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/amazon-simple-storage-service-connections-a7b660a.md b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/amazon-simple-storage-service-connections-a7b660a.md index c044659..733d4d4 100644 --- a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/amazon-simple-storage-service-connections-a7b660a.md +++ b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/amazon-simple-storage-service-connections-a7b660a.md @@ -96,7 +96,12 @@ Description -Enter the endpoint URL of the Amazon S3 server. The protocol prefix is not required. For example, `s3.amazonaws.com`. +Enter the endpoint URL of the Amazon S3 server, for example `s3.amazonaws.com`. The protocol prefix is not required. + +> ### Note: +> When using *Assume Role*, you must enter the regional endpoint, for example `s3.us-west-2.amazonaws.com`. + + diff --git a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/hadoop-distributed-file-system-connections-f9c3356.md b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/hadoop-distributed-file-system-connections-f9c3356.md index e99282e..b13d23d 100644 --- a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/hadoop-distributed-file-system-connections-f9c3356.md +++ b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/hadoop-distributed-file-system-connections-f9c3356.md @@ -51,7 +51,7 @@ Select *RPC* or *WEBHDFS* > ### Note: > Along with Remote Procedure Call \(RPC\), HDFS can also extend connections with WebHDFS. > -> If you select *RPC*, make sure that both the NameNode and the DataNodes are exposed and can be connected from SAP Datasphere. For more information, see [Finding SAP Datasphere IP addresses](https://help.sap.com/viewer/935116dd7c324355803d4b85809cec97/DEV_CURRENT/en-US/0934f7ed9a534e638299f53ab60866ae.html "Find externally facing IP addresses that for particular remote applications must be added to allowlists before you can to use connections to these remote applications.") :arrow_upper_right:. +> If you select *RPC*, make sure that both the NameNode and the DataNodes are exposed and can be connected from SAP Datasphere. For more information, see [Finding SAP Datasphere IP addresses](https://help.sap.com/viewer/935116dd7c324355803d4b85809cec97/DEV_CURRENT/en-US/0934f7ed9a534e638299f53ab60866ae.html "Find externally facing IP addresses and IDs that must be added to allowlists in particular remote applications before you can use connections to these remote applications.") :arrow_upper_right:. diff --git a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/microsoft-azure-data-lake-store-gen2-connections-cd06b3c.md b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/microsoft-azure-data-lake-store-gen2-connections-cd06b3c.md index 5ffe1f6..a28ff03 100644 --- a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/microsoft-azure-data-lake-store-gen2-connections-cd06b3c.md +++ b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/microsoft-azure-data-lake-store-gen2-connections-cd06b3c.md @@ -13,7 +13,7 @@ This topic contains the following sections: - [Supported Features](microsoft-azure-data-lake-store-gen2-connections-cd06b3c.md#loiocd06b3c5ab5147c0905e3fa8abd13eb1__ADL2_usage) - [Configuring Connection Properties](microsoft-azure-data-lake-store-gen2-connections-cd06b3c.md#loiocd06b3c5ab5147c0905e3fa8abd13eb1__connection_properties) -For information about the required prerequisites, see [Prepare Connectivity to Microsoft Azure Data Lake Store Gen2](https://help.sap.com/viewer/9f804b8efa8043539289f42f372c4862/cloud/en-US/6b3fd2c371284637bac465edb9cc9e50.html "To be able to successfully validate and use a connection to Microsoft Azure Data Lake Store Gen2 certain preparations have to be made.") :arrow_upper_right:. +For information about the required prerequisites, see [Prepare Connectivity to Microsoft Azure Data Lake Store Gen2](https://help.sap.com/viewer/935116dd7c324355803d4b85809cec97/DEV_CURRENT/en-US/6b3fd2c371284637bac465edb9cc9e50.html "To be able to successfully validate and use a connection to Microsoft Azure Data Lake Store Gen2 certain preparations have to be made.") :arrow_upper_right:. diff --git a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/oracle-connections-c73ae06.md b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/oracle-connections-c73ae06.md index 065698e..300d3a1 100644 --- a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/oracle-connections-c73ae06.md +++ b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/oracle-connections-c73ae06.md @@ -2,40 +2,73 @@ # Oracle Connections -Use an *Oracle* connection to access data from an Oracle database \(on-premise\). +Use the connection to connect to and access data from an Oracle database \(on-premise\). -## Prerequisites +This topic contains the following sections: -See: [Prepare Connectivity to Oracle](https://help.sap.com/viewer/935116dd7c324355803d4b85809cec97/DEV_CURRENT/en-US/9fca7c484e974429afc6570196303c35.html "To be able to successfully validate and use a connection to an Oracle database for remote tables or data flows, certain preparations have to be made.") :arrow_upper_right: +- [Supported Features](oracle-connections-c73ae06.md#loioc73ae0601d364f47830d339b6e86b7e8__Oracle_usage) +- [Configuring Connection Properties](oracle-connections-c73ae06.md#loioc73ae0601d364f47830d339b6e86b7e8__connection_properties) + +For information about the required prerequisites, see [Prepare Connectivity to Oracle](https://help.sap.com/viewer/935116dd7c324355803d4b85809cec97/DEV_CURRENT/en-US/9fca7c484e974429afc6570196303c35.html "To be able to successfully validate and use a connection to an Oracle database for remote tables or data flows, certain preparations have to be made.") :arrow_upper_right:. -## Using the Connection +## Supported Features -You can import database tables. The connection type supports the remote table as well as the data flow feature. -Supported data access methods for remote tables: + + + + + + + + + + + + + +
+ +Feature + + -- Remote -- Replication \(snapshot\) -- Replication \(real-time\) +Additional Information - > ### Note: - > SAP Datasphere uses trigger-based replication. For more information, see [Oracle Trigger-Based Replication](https://help.sap.com/viewer/7952ef28a6914997abc01745fef1b607/latest/en-US/0167d05537d84b0ea32979be85266b54.html) in the *SAP HANA Smart Data Integration and SAP HANA Smart Data Quality Installation and Configuration Guide*. +
+Remote Tables -For more information, see [Replicating Data and Monitoring Remote Tables](../Data-Integration-Monitor/replicating-data-and-monitoring-remote-tables-4dd95d7.md). + + +You can use remote tables imported from the connection either to access data directly live in the connected source \(federation\) or to copy the data into SAP Datasphere \(replication\). + +For remote tables, real-time replication is supported. For information about any constraints, see [Replicate Data Changes in Real-Time](../Data-Integration-Monitor/replicate-data-changes-in-real-time-441d327.md). > ### Note: -> Note that only databases available on the public internet can be consumed in data flows directly. To avoid this restriction, you can use a remote table as a source for your data flow. To do so, you can deploy a remote table by importing the data set you're interested in from the connection into an E/R model in your space, for example. When creating a data flow in the space, the remote table then is available from the *Repository* tab in the *Source Browser*. +> SAP Datasphere uses trigger-based replication. For more information, see [Oracle Trigger-Based Replication](https://help.sap.com/viewer/7952ef28a6914997abc01745fef1b607/latest/en-US/0167d05537d84b0ea32979be85266b54.html) in the *SAP HANA Smart Data Integration and SAP HANA Smart Data Quality Installation and Configuration Guide*. + + + +
+ +Data Flows + + +You can use the connection to add source objects to a data flow. + +
-
+ + ## Configuring Connection Properties @@ -123,6 +156,96 @@ Select a version. Supported versions are Oracle 12c, Oracle 18c, and Oracle 19c. +### Cloud Connector + +> ### Note: +> Cloud Connector is not required if your Oracle database is available on the public internet. + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +Property + + + +Description + +
+ +*Use Cloud Connector* + + + +\[optional\] Set to *true* if your source is an on-premise source and you want to use the connection for data flows. The default is *false*. + +
+ +\[if *Use Cloud Connector* = *true*\] *Location* + + + +\[optional\] Select a location ID. + +> ### Note: +> To select another location ID than the default location, *Connection.Read* privilege is required. The privilege is neither included in the *DW Integrator* nor in the *DW Space Administrator* role. If you need to select a location ID, ask your tenant administrator to either assign your user to a global role that is based on the *DW Administrator* role or to assign your user to a custom global role that includes the required *Connection.Read* privilege. + + + +
+ +\[if *Use Cloud Connector* = *true*\] *Virtual Destination* + + + +\[optional\] Select *Derive Virtual Host and Port from Connection Details* or *Enter Virtual Host and Port in Separate Fields*. + +If host and port entered in the connection details match the virtual host and port from the Cloud Connector configuration, you can select *Derive Virtual Host and Port from Connection Details* and don't need to enter the values manually. + +
+ +\[if *Virtual Destination* = *Enter Virtual Host and Port in Separate Fields*\] *Virtual Host* + + + +Enter the virtual host that you defined during Cloud Connector configuration. + +
+ +\[if *Virtual Destination* = *Enter Virtual Host and Port in Separate Fields*\] *Virtual Port* + + + +Enter the virtual port that you defined during Cloud Connector configuration. + +
+ + + ### Security diff --git a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-cloud-data-lake-files-connections-356e41e.md b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-cloud-data-lake-files-connections-356e41e.md index d35e68d..b06c707 100644 --- a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-cloud-data-lake-files-connections-356e41e.md +++ b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-cloud-data-lake-files-connections-356e41e.md @@ -35,7 +35,7 @@ Standalone SAP HANA Cloud, data lake is a standalone component in the SAP Busine > > - [Data Lake Connections](https://help.sap.com/docs/HANA_CLOUD/9ae9104a46f74a6583ce5182e7fb20cb/7e8ca90a9b4940d2930c36e92fbf6ba7.html) in the *SAP HANA Cloud* documentation > -> - [Finding SAP Datasphere IP addresses](https://help.sap.com/viewer/935116dd7c324355803d4b85809cec97/DEV_CURRENT/en-US/0934f7ed9a534e638299f53ab60866ae.html "Find externally facing IP addresses that for particular remote applications must be added to allowlists before you can to use connections to these remote applications.") :arrow_upper_right: +> - [Finding SAP Datasphere IP addresses](https://help.sap.com/viewer/935116dd7c324355803d4b85809cec97/DEV_CURRENT/en-US/0934f7ed9a534e638299f53ab60866ae.html "Find externally facing IP addresses and IDs that must be added to allowlists in particular remote applications before you can use connections to these remote applications.") :arrow_upper_right: diff --git a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-cloud-data-lake-relational-engine-connections-40763e2.md b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-cloud-data-lake-relational-engine-connections-40763e2.md index 3866b32..67a6598 100644 --- a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-cloud-data-lake-relational-engine-connections-40763e2.md +++ b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-cloud-data-lake-relational-engine-connections-40763e2.md @@ -32,7 +32,7 @@ Supported data access methods for remote tables: > > - [Data Lake Connections](https://help.sap.com/docs/HANA_CLOUD/9ae9104a46f74a6583ce5182e7fb20cb/7e8ca90a9b4940d2930c36e92fbf6ba7.html) in the *SAP HANA Cloud* documentation > -> - [Finding SAP Datasphere IP addresses](https://help.sap.com/viewer/935116dd7c324355803d4b85809cec97/DEV_CURRENT/en-US/0934f7ed9a534e638299f53ab60866ae.html "Find externally facing IP addresses that for particular remote applications must be added to allowlists before you can to use connections to these remote applications.") :arrow_upper_right: +> - [Finding SAP Datasphere IP addresses](https://help.sap.com/viewer/935116dd7c324355803d4b85809cec97/DEV_CURRENT/en-US/0934f7ed9a534e638299f53ab60866ae.html "Find externally facing IP addresses and IDs that must be added to allowlists in particular remote applications before you can use connections to these remote applications.") :arrow_upper_right: diff --git a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-connections-e6b63f1.md b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-connections-e6b63f1.md index 77f300a..440d830 100644 --- a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-connections-e6b63f1.md +++ b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-hana-connections-e6b63f1.md @@ -139,9 +139,7 @@ Enter the fully qualified host name or IP address on which the remote SAP HANA s -Enter the SQL port number of the remote SAP HANA server. - -You can find the SQL port in the list of service details in the SAP HANA Cockpit. For more information, see [Service Details](https://help.sap.com/viewer/afa922439b204e9caf22c78b6b69e4f2/latest/en-US/1d4b4fe9ee8b448a9977bfcc1fc55248.html?q=SQL%20port) in the *SAP HANA Administration with SAP HANA Cockpit* documentation. +Enter the SQL port number of the SAP HANA Cloud instance: 443. @@ -457,9 +455,9 @@ Enter the fully qualified host name or IP address on which the remote SAP HANA s -Enter the SQL port number of the remote SAP HANA server. +Enter the TCP SQL port number of the remote SAP HANA server. -You can find the SQL port in the list of service details in the SAP HANA Cockpit. For more information, see [Service Details](https://help.sap.com/viewer/afa922439b204e9caf22c78b6b69e4f2/latest/en-US/1d4b4fe9ee8b448a9977bfcc1fc55248.html?q=SQL%20port) in the *SAP HANA Administration with SAP HANA Cockpit* documentation. +For more information about port assignment for SAP HANA, see [Port Assignment in Tenant Databases](https://help.sap.com/docs/SAP_HANA_PLATFORM/6b94445c94ae495c83a19646e7c3fd56/440f6efe693d4b82ade2d8b182eb1efb.html) in the *SAP HANA Platform* documentation. diff --git a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-s-4hana-cloud-connections-a98e5ff.md b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-s-4hana-cloud-connections-a98e5ff.md index b47f06c..b5da9ac 100644 --- a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-s-4hana-cloud-connections-a98e5ff.md +++ b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-s-4hana-cloud-connections-a98e5ff.md @@ -47,10 +47,19 @@ You can use remote tables imported from the connection either to access data dir You can access the following data: -- ABAP CDS-Views exposed as OData services for data extraction \(access via Cloud Data Integration\) +- for federation and replication: ABAP CDS Views exposed as OData services for data extraction \(access via Cloud Data Integration\) + + For remote tables, real-time replication is supported. For information about any constraints, see [Replicate Data Changes in Real-Time](../Data-Integration-Monitor/replicate-data-changes-in-real-time-441d327.md). + +- for federation: standard and custom CDS view entities that are exposed using the SQL service from SAP S/4HANA Cloud + + For more information, see [Using ABAP SQL Services for Accessing Data from SAP S/4HANA Cloud](https://help.sap.com/viewer/9f804b8efa8043539289f42f372c4862/cloud/en-US/ef2b2238154f4cd78a08df360447c1d5.html "The ABAP SQL service provides SQL-level access to published CDS view entities for SAP Datasphere. You can use the service to replicate data with replication flows or to federate data with remote tables.") :arrow_upper_right:. + + > ### Note: + > On the *Sources* tab of the remote-table-related Data Builder editors in SAP Datasphere, the service binding name from the *SQL\_SCHEMA* authorization field is visible as \(virtual\) schema. + -For remote tables, real-time replication is supported. For information about any constraints, see [Replicate Data Changes in Real-Time](../Data-Integration-Monitor/replicate-data-changes-in-real-time-441d327.md). @@ -373,6 +382,102 @@ Enter the user password. +### Remote Tables + + + + + + + + + + + + + + + + + + + + + + + +
+ +Property + + + +Description + +
+ +*Data Provisioning Option* + + + +Select the middleware to use when connecting to and accessing SAP S/4HANA Cloud: + +- *None* \(default\): if you don't want to use remote tables. + +- *Data Provisioning Agent*: if you want to use federation and replication via SAP HANA smart data integration CloudDataIntegrationAdapter. + +- *Direct*: if you want to federate data from the source objects of the connection and access them remotely in the source system using the ABAP SQL service. + + > ### Note: + > This option is supported for authentication type *User Name And Password*. + + + + +
+ +\[if *Data Provisioning Option* = *Data Provisioning Agent* or *Direct*\] *Data Access* + + + +\[read-only\] Displays how data from source objects can be accessed: + +- *Remote and Replication*: if you selected data provisioning option *Data Provisioning Agent*. + +- *Remote Only*: if you selected data provisioning option *Direct*. + + + + +
+ +\[if *Data Provisioning Option* = *Data Provisioning Agent*\] *Data Provisioning Agent* + + + +Select an agent from the list of agents configured for SAP Datasphere. + +
+ +\[if *Data Provisioning Option* = *Direct*\] *Type Map* + + + +Specify how ABAP data types are mapped to SAP HANA data types: + +- *native*: no conversion of ABAP data types + +- *semantic*: conversion of the ABAP data types to an ABAP-independent data format + +- *semanticDatsTimsAsWchar*: like *semantic*, except that the ABAP types DATS and TIMS are mapped to the ODBC type SQL\_WCHAR to allow for lossless conversion of date literals and time literals + + + + +
+ + + ### Features @@ -397,7 +502,7 @@ Description -To enable *Remote Tables*, select a Data Provisioning Agent. +To enable*Remote Tables*, complete the connection properties in the *Remote Tables* section. diff --git a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-successfactors-connections-39df020.md b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-successfactors-connections-39df020.md index 914a6d9..e2c54b2 100644 --- a/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-successfactors-connections-39df020.md +++ b/docs/Integrating-data-and-managing-spaces/Integrating-Data-Via-Connections/sap-successfactors-connections-39df020.md @@ -51,6 +51,8 @@ You can use remote tables imported from the connection either to access data dir > Please note that since the *Generic OData* connection type doesn’t use snapshot-based pagination at all, we do not recommend to use it to federate or replicate SAP SuccessFactors data with remote tables. > > - When replicating employee-related data, independently of their validity not only current data is replicated but all data including historical data. +> +> - Picklists used in SAP SuccessFactors are not available in remote tables because they are exposed as navigation properties and not as exposed properties of the entity. For more information on picklists, see SAP Note [2088679](https://me.sap.com/notes/2088679). diff --git a/docs/Integrating-data-and-managing-spaces/Transporting-Content-Between-Tenants/transporting-your-content-through-sap-cloud-transport-management-0538398.md b/docs/Integrating-data-and-managing-spaces/Transporting-Content-Between-Tenants/transporting-your-content-through-sap-cloud-transport-management-0538398.md index 86c8b55..725527d 100644 --- a/docs/Integrating-data-and-managing-spaces/Transporting-Content-Between-Tenants/transporting-your-content-through-sap-cloud-transport-management-0538398.md +++ b/docs/Integrating-data-and-managing-spaces/Transporting-Content-Between-Tenants/transporting-your-content-through-sap-cloud-transport-management-0538398.md @@ -74,7 +74,7 @@ Once the import is completed, you see the content of the package available in it ## Troubleshooting -In case of import failure, you can access the logs to investigate the error by adding this endpoint to the target tenant URL: +In case of import failure, you can access the logs to investigate the error by adding this endpoint to the target tenant URL in your browser: -`/contentmanager/v4/jobs/{ProcessID}?brief=false` +`{TenantURL}/contentmanager/v4/jobs/{ProcessID}?brief=false` diff --git a/docs/Integrating-data-and-managing-spaces/lock-or-unlock-your-space-c05b6a6.md b/docs/Integrating-data-and-managing-spaces/lock-or-unlock-your-space-c05b6a6.md index 6ef8754..885a8a2 100644 --- a/docs/Integrating-data-and-managing-spaces/lock-or-unlock-your-space-c05b6a6.md +++ b/docs/Integrating-data-and-managing-spaces/lock-or-unlock-your-space-c05b6a6.md @@ -2,7 +2,7 @@ # Lock or Unlock Your Space -If a space exceeds its assigned storage or if the audit logs enabled in the space consume too much disk storage, the space is automatically locked. +When a space exceeds its assigned storage or when the audit logs enabled in the space consume too much disk storage, the space is automatically locked after 60 minutes if you do not free up space. If you want to put a space on hold and you are a space administrator, you can manually lock it.