Skip to content

Commit d62b8f3

Browse files
author
Deepak Sahu
committed
ai-core: typo & docker secret URL update
- update Docker resgistry URL support is `https://index.docker.io` . Other URL dont work.
1 parent a1a5796 commit d62b8f3

5 files changed

Lines changed: 8 additions & 8 deletions

File tree

tutorials/ai-core-code/ai-core-code.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -246,7 +246,7 @@ print(response.__dict__)
246246
[OPTION END]
247247

248248
1. **Name**: Enter `credstutorialrepo`. This is becomes an identifier for your Docker credentials within SAP AI Core. This value is your docker registry secret.
249-
2. **URL**: If you have used your organization's Docker registry then use its URL, otherwise, enter `docker.io`.
249+
2. **URL**: If you have used your organization's Docker registry then use its URL, otherwise, enter `https://index.docker.io`.
250250
3. **Username**: Your Docker username.
251251
4. **Access Token**: The access token generated previously, in the Docker account settings.
252252

16.8 KB
Loading
415 Bytes
Loading
12.9 KB
Loading

tutorials/ai-core-data/ai-core-data.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ Your code reads the data file `train.csv` from the location `/app/data`. This da
7474

7575
!![image](img/code-main.png)
7676

77-
Create file `requirement.txt` as shown below. Here, if you don't specify a particular version like `pandas` then the latest version of the package will be fetched automatically.
77+
Create file `requirements.txt` as shown below. Here, if you don't specify a particular version like `pandas` then the latest version of the package will be fetched automatically.
7878

7979
```TEXT
8080
sklearn==0.0
@@ -110,7 +110,7 @@ RUN chgrp -R 65534 /app && \
110110

111111
!![image](img/code-docker.png)
112112

113-
> **IMPORTANT** Your `Dockerfile` createS empty folders to store your datasets and models (example above `/app/data` and `/app/model/` ). Contents from cloud storage will be copied to and from these folders later. If you place any contents in these folders during time of Docker image building, then the contents will be overwritten.
113+
> **IMPORTANT** Your `Dockerfile` creates empty folders to store your datasets and models (example above `/app/data` and `/app/model/` ). Contents from cloud storage will be copied to and from these folders later. If you place any contents in these folders during time of Docker image building, then the contents will be overwritten.
114114
115115
Build and upload your Docker image to Docker repository, using the following code in the terminal.
116116

@@ -124,7 +124,7 @@ docker push docker.io/<YOUR_DOCKER_USERNAME>/house-price:03
124124

125125
[ACCORDION-BEGIN [Step 2: ](Create placeholders for datasets in workflows)]
126126

127-
Create a pipeline (yaml file) named `house-price-train.yaml` in your GitHub repository. You may use the existing GitHub path which is already tracked (auto synced) by your application of SAP AI Core.
127+
Create a pipeline (YAML file) named `house-price-train.yaml` in your GitHub repository. You may use the existing GitHub path which is already tracked (auto synced) by your application of SAP AI Core.
128128

129129
```YAML
130130
apiVersion: argoproj.io/v1alpha1
@@ -696,11 +696,11 @@ Copy the artifact ID of the January dataset. You will use this value in the plac
696696

697697
[OPTION BEGIN [SAP AI Launchpad]]
698698

699-
Click through **ML Operations** > **Configuration** > **Create**. Enter the following details as shown in the image below. Click **Next**
699+
Click through **ML Operations** > **Configuration** > **Create**. Enter the following details as shown in the image below. Click **Next**.
700700

701701
!![image](img/ail/config-1.png)
702702

703-
The field for `DT_MAX_DEPTH` allows your to use the configuration to pass values to placeholders of hyper-parameters that your prepared earlier in your workflows. In this case, type `3`. Click **Next**.
703+
The field for `DT_MAX_DEPTH` allows you to use the configuration to pass values to placeholders of hyper-parameters that you prepared earlier in your workflows. In this case, type `3`. Click **Next**.
704704

705705
!![image](img/ail/config-2.png)
706706

@@ -749,7 +749,7 @@ Use the artifact ID of the `jan` dataset and the placeholder names to create a c
749749

750750
[OPTION BEGIN [SAP AI Core SDK]]
751751

752-
Paste and edit the code snippet. The key value pair for `DT_MAX_DEPTH` allows your to use the configuration to pass values to placeholders of hyper-parameters that your prepared earlier in your workflows. In this case, type `3`.
752+
Paste and edit the code snippet. The key value pair for `DT_MAX_DEPTH` allows your to use the configuration to pass values to placeholders of hyper-parameters that your prepared earlier in your workflows. In this case, type `3`. You should locate your `jan` dataset artifact ID by listing all artifacts and use the relevant ID.
753753

754754
```PYTHON
755755
from ai_api_client_sdk.models.parameter_binding import ParameterBinding
@@ -1029,7 +1029,7 @@ Paste and edit the snippet below. You should locate your `feb` dataset artifact
10291029

10301030
[OPTION BEGIN [SAP AI Core SDK]]
10311031

1032-
Paste and edit the snippet below. You should locate your `feb` dataset artifact ID by listing all artifacts and using the relevant ID.
1032+
Paste and edit the snippet below. You should locate your `feb` dataset artifact ID by listing all artifacts and use the relevant ID.
10331033

10341034
```PYTHON
10351035
from ai_api_client_sdk.models.parameter_binding import ParameterBinding

0 commit comments

Comments
 (0)