Skip to content

Commit 6e131b7

Browse files
authored
Merge pull request #821 from qingqing01/fix_dead_link
Fix dead links in quick start after reorganzing documentaion directory
2 parents 3b1acd2 + 4279a23 commit 6e131b7

File tree

1 file changed

+8
-7
lines changed

1 file changed

+8
-7
lines changed

doc/tutorials/quick_start/index_en.md

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ This tutorial will teach the basics of deep learning (DL), including how to impl
1212

1313
To get started, please install PaddlePaddle on your computer. Throughout this tutorial, you will learn by implementing different DL models for text classification.
1414

15-
To install PaddlePaddle, please follow the instructions here: <a href = "../../build/index.html" >Build and Install</a>.
15+
To install PaddlePaddle, please follow the instructions here: <a href = "../../getstarted/build_and_install/index_en.html" >Build and Install</a>.
1616

1717
## Overview
1818
For the first step, you will use PaddlePaddle to build a **text classification** system. For example, suppose you run an e-commence website, and you want to analyze the sentiment of user reviews to evaluate product quality.
@@ -156,14 +156,14 @@ define_py_data_sources2(train_list='data/train.list',
156156
obj="process",
157157
args={"dictionary": word_dict})
158158
```
159-
You can refer to the following link for more detailed examples and data formats: <a href = "../../ui/data_provider/pydataprovider2.html">PyDataProvider2</a>.
159+
You can refer to the following link for more detailed examples and data formats: <a href = "../../api/data_provider/pydataprovider2_en.html">PyDataProvider2</a>.
160160

161161
## Network Architecture
162162
You will describe four kinds of network architectures in this section.
163163
<center> ![](./PipelineNetwork_en.jpg) </center>
164164

165165
First, you will build a logistic regression model. Later, you will also get chance to build other more powerful network architectures.
166-
For more detailed documentation, you could refer to: <a href = "../../ui/api/trainer_config_helpers/layers_index.html">Layer documentation</a>All configuration files are in `demo/quick_start` directory.
166+
For more detailed documentation, you could refer to: <a href = "../../api/trainer_config_helpers/layers.html">layer documentation</a>. All configuration files are in `demo/quick_start` directory.
167167

168168
### Logistic Regression
169169
The architecture is illustrated in the following picture:
@@ -366,7 +366,7 @@ You can use single layer LSTM model with Dropout for our text classification pro
366366
<br>
367367

368368
## Optimization Algorithm
369-
<a href = "../../ui/api/trainer_config_helpers/optimizers.html">Optimization algorithms</a> include Momentum, RMSProp, AdaDelta, AdaGrad, Adam, and Adamax. You can use Adam optimization method here, with L2 regularization and gradient clipping, because Adam has been proved to work very well for training recurrent neural network.
369+
<a href = "../../api/trainer_config_helpers/optimizers.html">Optimization algorithms</a> include Momentum, RMSProp, AdaDelta, AdaGrad, Adam, and Adamax. You can use Adam optimization method here, with L2 regularization and gradient clipping, because Adam has been proved to work very well for training recurrent neural network.
370370

371371
```python
372372
settings(batch_size=128,
@@ -391,7 +391,8 @@ paddle train \
391391
--use_gpu=false
392392
```
393393

394-
If you want to install the remote training platform, which enables distributed training on clusters, follow the instructions here: <a href = "../../cluster/index.html">Platform</a> documentation. We do not provide examples on how to train on clusters. Please refer to other demos or platform training documentation for mode details on training on clusters.
394+
We do not provide examples on how to train on clusters here. If you want to train on clusters, please follow the <a href = "../../howto/cluster/cluster_train_en.html">distributed training</a> documentation or other demos for more details.
395+
395396
## Inference
396397
You can use the trained model to perform prediction on the dataset with no labels. You can also evaluate the model on dataset with labels to obtain its test accuracy.
397398
<center> ![](./PipelineTest_en.png) </center>
@@ -406,7 +407,7 @@ paddle train \
406407
--init_model_path=./output/pass-0000x
407408
```
408409

409-
We will give an example of performing prediction using Recurrent model on a dataset with no labels. You can refer to: <a href = "../../ui/predict/swig_py_paddle_en.html">Python Prediction API</a> tutorial,or other <a href = "../../demo/index.html">demo</a> for the prediction process using Python. You can also use the following script for inference or evaluation.
410+
We will give an example of performing prediction using Recurrent model on a dataset with no labels. You can refer to <a href = "../../api/predict/swig_py_paddle_en.html">Python Prediction API</a> tutorial,or other <a href = "../../tutorials/index_en.html">demo</a> for the prediction process using Python. You can also use the following script for inference or evaluation.
410411

411412
inference script (predict.sh):
412413

@@ -508,7 +509,7 @@ The scripts of data downloading, network configurations, and training scrips are
508509
* \--config_args:Other configuration arguments.
509510
* \--init_model_path:The path of the initial model parameter.
510511

511-
By default, the trainer will save model every pass. You can also specify `saving_period_by_batches` to set the frequency of batch saving. You can use `show_parameter_stats_period` to print the statistics of the parameters, which are very useful for tuning parameters. Other command line arguments can be found in <a href = "../../ui/index.html#command-line-argument">command line argument documentation</a>。
512+
By default, the trainer will save model every pass. You can also specify `saving_period_by_batches` to set the frequency of batch saving. You can use `show_parameter_stats_period` to print the statistics of the parameters, which are very useful for tuning parameters. Other command line arguments can be found in <a href = "../../howto/cmd_parameter/index_en.html">command line argument documentation</a>。
512513

513514
### Log
514515

0 commit comments

Comments
 (0)