[
https://issues.apache.org/jira/browse/BEAM-13582?focusedWorklogId=759506&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-759506
]
ASF GitHub Bot logged work on BEAM-13582:
-----------------------------------------
Author: ASF GitHub Bot
Created on: 20/Apr/22 19:10
Start Date: 20/Apr/22 19:10
Worklog Time Spent: 10m
Work Description: tvalentyn commented on code in PR #17300:
URL: https://github.com/apache/beam/pull/17300#discussion_r854428337
##########
website/www/site/content/en/documentation/resources/learning-resources.md:
##########
@@ -97,8 +97,7 @@ If you have additional material that you would like to see
here, please let us k
### Python
-* **[Python Qwik
Start](https://qwiklabs.com/focuses/1100?locale=en&parent=catalog)** (30m) -
Run a word count pipeline on the Dataflow runner.
-* **[NDVI from Landsat
Images](https://qwiklabs.com/focuses/1849?locale=en&parent=catalog)** (45m) -
Process Landsat satellite data in a distributed environment to compute the
[Normalized Difference Vegetation
Index](https://en.wikipedia.org/wiki/Normalized_difference_vegetation_index)
(NDVI).
+* **[Python Qwik
Start](https://www.qwiklabs.com/focuses/1098?parent=catalog)** (30m) - Run a
word count pipeline on the Dataflow runner.
Review Comment:
not sure what's the difference in the urls for the first link; both
...1098... and ...1100... work for me (after signing in with Qwiklabs). the
removed `NDVI from Landsat Images` link did not work, indeed.
##########
website/www/site/content/en/documentation/sdks/go.md:
##########
@@ -28,7 +28,7 @@ This means worker binaries may need to be
[cross-compiled](/documentation/sdks/g
Get started with the [Beam Go SDK quickstart](/get-started/quickstart-go) to
set up your development environment and run an example pipeline. Then, read
through the [Beam programming guide](/documentation/programming-guide) to learn
the basic concepts that apply to all SDKs in Beam.
-See the [godoc](https://pkg.go.dev/github.com/apache/beam/sdks/v2/go/pkg/beam)
for more detailed information.
+See the [godoc](https://pkg.go.dev/github.com/apache/beam/sdks/go/pkg/beam)
for more detailed information.
Review Comment:
What's the reason for this change? Both links work. I think v2 refers to
documentation for newer SDKs and will be used going forward.
##########
website/www/site/content/en/blog/review-input-streaming-connectors.md:
##########
@@ -184,15 +184,15 @@ and <a
href="https://spark.apache.org/docs/latest/api/python/pyspark.streaming.h
</td>
<td>N/A
</td>
- <td><a
href="https://spark.apache.org/docs/latest/api/python/pyspark.streaming.html#pyspark.streaming.kafka.KafkaUtils">KafkaUtils</a>
+ <td><a
href="https://spark.apache.org/docs/latest/streaming-kafka-0-10-integration.html">KafkaUtils</a>
Review Comment:
this is an old blog and the info is outdated.
@rszper you could perhaps reference older version of pyspark to preserve the
original narrative.
https://spark.apache.org/docs/2.4.8/api/python/pyspark.streaming.html#pyspark.streaming.kafka.KafkaUtils
##########
website/www/site/content/en/documentation/runtime/environments.md:
##########
@@ -102,7 +102,7 @@ This method requires building image artifacts from Beam
source. For additional i
git checkout origin/release-$BEAM_SDK_VERSION
```
-2. Customize the `Dockerfile` for a given language, typically
`sdks/<language>/container/Dockerfile` directory (e.g. the [Dockerfile for
Python](https://github.com/apache/beam/blob/master/sdks/python/container/Dockerfile).
If you're adding dependencies from [PyPI](https://pypi.org/), use
[`base_image_requirements.txt`](https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt)
instead.
+2. Customize the `Dockerfile` for a given language, typically
`sdks/<language>/container/Dockerfile` directory (e.g. the [Dockerfile for
Python](https://github.com/apache/beam/blob/master/sdks/python/container/Dockerfile).
If you're adding dependencies from [PyPI](https://pypi.org/), use
[`base_image_requirements.txt`](https://github.com/apache/beam/blob/master/sdks/python/container/py38/base_image_requirements.txt)
instead.
Review Comment:
Let's remove If you're adding ... sentence.
##########
website/www/site/content/en/blog/review-input-streaming-connectors.md:
##########
@@ -184,15 +184,15 @@ and <a
href="https://spark.apache.org/docs/latest/api/python/pyspark.streaming.h
</td>
<td>N/A
</td>
- <td><a
href="https://spark.apache.org/docs/latest/api/python/pyspark.streaming.html#pyspark.streaming.kafka.KafkaUtils">KafkaUtils</a>
+ <td><a
href="https://spark.apache.org/docs/latest/streaming-kafka-0-10-integration.html">KafkaUtils</a>
</td>
</tr>
<tr>
<td>Kinesis
</td>
<td>N/A
</td>
- <td><a
href="https://spark.apache.org/docs/latest/api/python/pyspark.streaming.html#module-pyspark.streaming.kinesis">KinesisUtils</a>
+ <td><a
href="https://spark.apache.org/docs/latest/api/python/reference/api/pyspark.streaming.kinesis.KinesisUtils.createStream.html">KinesisUtils</a>
Review Comment:
ditto
Issue Time Tracking
-------------------
Worklog Id: (was: 759506)
Time Spent: 1h 40m (was: 1.5h)
> Beam website precommit mentions broken links, but passes.
> ---------------------------------------------------------
>
> Key: BEAM-13582
> URL: https://issues.apache.org/jira/browse/BEAM-13582
> Project: Beam
> Issue Type: Bug
> Components: website
> Reporter: Valentyn Tymofieiev
> Assignee: Rebecca Szper
> Priority: P1
> Labels: starter
> Time Spent: 1h 40m
> Remaining Estimate: 0h
>
> There may be two AIs here - fix broken links, consider making the check fail
> when broken links are found.
> Sample log output:
> 15:46:51 Found 26 invalid links:
> 15:46:51 http://jstorm.io/
> 15:46:51 http://jstorm.io/QuickStart/Deploy/index.html
> 15:46:51
> https://beam.apache.org/releases/pydoc/2.34.0/apache_beam.testing.util.html#apache_beam.testing.util.assert_that
> 15:46:51 https://drive.google.com/open?id=1EbijvZKpkWwWyMryLY9sJfyZzZk1k44v
> 15:46:51
> https://github.com/apache/beam/blob/master/.test-infra/kubernetes/hadoop/LargeITCluster/pkb-config.yml
> 15:46:51
> https://github.com/apache/beam/blob/master/.test-infra/kubernetes/hadoop/SmallITCluster/pkb-config.yml
> 15:46:51
> https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt
> 15:46:51
> https://github.com/googleapis/google-cloud-python/blob/master/bigquery/google/cloud/bigquery/table.py#L153
> 15:46:51 https://issues.apache/jira/browse/BEAM-11929
> 15:46:51 https://issues.apache/jira/browse/BEAM-12071
> 15:46:51 https://lists.apache.org/[email protected]
> 15:46:51 https://lists.apache.org/[email protected]
> 15:46:51
> https://pkg.go.dev/github.com/apache/beam/sdks/v2/go/pkg/beam/core/graph/coders
> 15:46:51 https://qwiklabs.com/focuses/1849?locale=en&parent=catalog
> 15:46:51 https://reporter.apache.org/addrelease.html?beam
> 15:46:51
> https://repository.apache.org/content/repositories/orgapachebeam-NNNN/
> 15:46:51
> https://repository.apache.org/content/repositories/orgapachebeam-NNNN/)
> 15:46:51 https://s.apache.org/io-test-dashboards
> 15:46:51 https://spark.apache.org/docs/latest/api/python/pyspark.html
> 15:46:51
> https://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.streaming.DataStreamReader
> 15:46:51
> https://spark.apache.org/docs/latest/api/python/pyspark.streaming.html#module-pyspark.streaming.kinesis
> 15:46:51
> https://spark.apache.org/docs/latest/api/python/pyspark.streaming.html#pyspark.streaming.StreamingContext.textFileStream
> 15:46:51
> https://spark.apache.org/docs/latest/api/python/pyspark.streaming.html#pyspark.streaming.kafka.KafkaUtils
> 15:46:51 https://store-beam.myshopify.com/
> 15:46:51 https://www.artstation.com/jbruno
> 15:46:51 https://www.linkedin.com/company/beam-summit/
> cc: [~pcoet]
--
This message was sent by Atlassian Jira
(v8.20.7#820007)