Serverless Data Processing with Dataflow - Advanced Streaming Analytics Pipeline with Dataflow (Python) Reviews

4994 reviews

NOUREDDINE E. · Reviewed about 4 years ago

Errori Python. segnalata issue in github

Fabrizio P. · Reviewed about 4 years ago

Errori Python irrisolvibili. (df-env) (base) jupyter@theia-20220321-123339:~/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab$ python3 streaming_minute_traffic_pipeline.py --project=${PROJECT_ID} --region=${REGION} --staging_location=${PIPELINE_FOLDER}/staging --temp_location=${PIPELINE_FOLDER}/temp --runner=${RUNNER} --input_topic=${PUBSUB_TOPIC} --window_duration=${WINDOW_DURATION} --allowed_lateness=${ALLOWED_LATENESS} --table_name=${OUTPUT_TABLE_NAME} --dead_letter_bucket=${DEADLETTER_BUCKET} streaming_minute_traffic_pipeline.py:116: FutureWarning: WriteToFiles is experimental. max_writers_per_bundle=0) /home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/io/fileio.py:550: BeamDeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported p.options.view_as(GoogleCloudOptions).temp_location or Traceback (most recent call last): File "streaming_minute_traffic_pipeline.py", line 140, in <module> run() File "streaming_minute_traffic_pipeline.py", line 116, in run max_writers_per_bundle=0) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/pvalue.py", line 137, in __or__ return self.pipeline.apply(ptransform, self) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/pipeline.py", line 652, in apply transform.transform, pvalueish, label or transform.label) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/pipeline.py", line 662, in apply return self.apply(transform, pvalueish) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/pipeline.py", line 708, in apply pvalueish_result = self.runner.apply(transform, pvalueish, self._options) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/runners/dataflow/dataflow_runner.py", line 141, in apply return super().apply(transform, input, options) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/runners/runner.py", line 185, in apply return m(transform, input, options) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/runners/runner.py", line 215, in apply_PTransform return transform.expand(input) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/io/fileio.py", line 577, in expand | beam.ParDo( File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/pvalue.py", line 137, in __or__ return self.pipeline.apply(ptransform, self) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/pipeline.py", line 652, in apply transform.transform, pvalueish, label or transform.label) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/pipeline.py", line 662, in apply return self.apply(transform, pvalueish) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/pipeline.py", line 708, in apply pvalueish_result = self.runner.apply(transform, pvalueish, self._options) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/runners/dataflow/dataflow_runner.py", line 141, in apply return super().apply(transform, input, options) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/runners/runner.py", line 185, in apply return m(transform, input, options) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/runners/dataflow/dataflow_runner.py", line 830, in apply_GroupByKey return transform.expand(pcoll) File "/home/jupyter/project/training-data-analyst/quests/dataflow_python/7_Advanced_Streaming_Analytics/lab/df-env/lib/python3.7/site-packages/apache_beam/transforms/core.py", line 2609, in expand raise ValueError(msg) ValueError: GroupRecordsByDestinationAndShard: Unsafe trigger: `AfterProcessingTime(delay=120)` may lose data. Reason: MAY_FINISH. This can be overriden with the --allow_unsafe_triggers flag. (df-env) (base)

Fabrizio P. · Reviewed about 4 years ago

Ray K. · Reviewed about 4 years ago

Edyta K. · Reviewed about 4 years ago

Rubén L. · Reviewed about 4 years ago

Ehsan U. · Reviewed about 4 years ago

Rubén L. · Reviewed about 4 years ago

Jacob V. · Reviewed about 4 years ago

Richard K. · Reviewed about 4 years ago

Irene K. · Reviewed about 4 years ago

Excellent level of detail and description.

Tristan T. · Reviewed about 4 years ago

shaik T. · Reviewed about 4 years ago

Marek D. · Reviewed about 4 years ago

Python run command for the pipeline gives an error. Requires an extra flag to run

Berkay B. · Reviewed about 4 years ago

Brady B. · Reviewed about 4 years ago

Alexandre S. · Reviewed about 4 years ago

Abhishek C. · Reviewed about 4 years ago

few stuff to modify to have the results but ok

Charles M. · Reviewed about 4 years ago

Zelmar M. · Reviewed about 4 years ago

There are a couple of problems with this lab. First, when running the pipeline I got the following error: ValueError: GroupRecordsByDestinationAndShard: Unsafe trigger: `AfterProcessingTime(delay=120)` may lose data. Reason: MAY_FINISH. This can be overriden with the --allow_unsafe_triggers flag. I was able to fix this by changing this: trigger=AfterProcessingTime(120) to this: trigger=Repeatedly(AfterProcessingTime(120)) Second, we are asked to add the following to ConvertToCommonLogFn: yield beam.pvalue.TaggedOutput('unparsed_row', element.decode('utf-8')) However the lab code already includes the following in the previous line: element = element.decode('utf-8') If this isn't removed, we get the following error due to double-decoding: AttributeError: 'str' object has no attribute 'decode' [while running 'ParseJson/ParDo(ConvertToCommonLogFn)-ptransform-81']

Robert L. · Reviewed about 4 years ago

Hanh T. · Reviewed about 4 years ago

Edward U. · Reviewed about 4 years ago

Jack R. · Reviewed about 4 years ago

Thiago M. · Reviewed about 4 years ago

We do not ensure the published reviews originate from consumers who have purchased or used the products. Reviews are not verified by Google.