github.com/apache/beam/sdks/v2@v2.48.2/python/apache_beam/io/hdfs_integration_test/Dockerfile (about) 1 ############################################################################### 2 # Licensed to the Apache Software Foundation (ASF) under one 3 # or more contributor license agreements. See the NOTICE file 4 # distributed with this work for additional information 5 # regarding copyright ownership. The ASF licenses this file 6 # to you under the Apache License, Version 2.0 (the 7 # "License"); you may not use this file except in compliance 8 # with the License. You may obtain a copy of the License at 9 # 10 # http://www.apache.org/licenses/LICENSE-2.0 11 # 12 # Unless required by applicable law or agreed to in writing, software 13 # distributed under the License is distributed on an "AS IS" BASIS, 14 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 15 # See the License for the specific language governing permissions and 16 # limitations under the License. 17 ############################################################################### 18 19 # This image contains a Python SDK build and dependencies. 20 # By default it runs wordcount against a locally accessible HDFS service. 21 # See hdfs_integration_test.sh for example usage. 22 ARG BASE_IMAGE 23 FROM $BASE_IMAGE 24 25 WORKDIR /app 26 ENV HDFSCLI_CONFIG /app/sdks/python/apache_beam/io/hdfs_integration_test/hdfscli.cfg 27 28 # Add Beam SDK sources. 29 COPY sdks/python /app/sdks/python 30 COPY model /app/model 31 32 # This step should look like setupVirtualenv minus virtualenv creation. 33 RUN pip install --no-cache-dir tox==3.11.1 -r sdks/python/build-requirements.txt 34 35 # Run wordcount, and write results to HDFS. 36 CMD cd sdks/python && tox -e hdfs_integration_test