github.com/apache/beam/sdks/v2@v2.48.2/python/apache_beam/io/hdfs_integration_test/Dockerfile (about)

     1  ###############################################################################
     2  #  Licensed to the Apache Software Foundation (ASF) under one
     3  #  or more contributor license agreements.  See the NOTICE file
     4  #  distributed with this work for additional information
     5  #  regarding copyright ownership.  The ASF licenses this file
     6  #  to you under the Apache License, Version 2.0 (the
     7  #  "License"); you may not use this file except in compliance
     8  #  with the License.  You may obtain a copy of the License at
     9  #
    10  #      http://www.apache.org/licenses/LICENSE-2.0
    11  #
    12  #  Unless required by applicable law or agreed to in writing, software
    13  #  distributed under the License is distributed on an "AS IS" BASIS,
    14  #  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    15  #  See the License for the specific language governing permissions and
    16  # limitations under the License.
    17  ###############################################################################
    18  
    19  # This image contains a Python SDK build and dependencies.
    20  # By default it runs wordcount against a locally accessible HDFS service.
    21  # See hdfs_integration_test.sh for example usage.
    22  ARG BASE_IMAGE
    23  FROM $BASE_IMAGE
    24  
    25  WORKDIR /app
    26  ENV HDFSCLI_CONFIG /app/sdks/python/apache_beam/io/hdfs_integration_test/hdfscli.cfg
    27  
    28  # Add Beam SDK sources.
    29  COPY sdks/python /app/sdks/python
    30  COPY model /app/model
    31  
    32  # This step should look like setupVirtualenv minus virtualenv creation.
    33  RUN pip install --no-cache-dir tox==3.11.1 -r sdks/python/build-requirements.txt
    34  
    35  # Run wordcount, and write results to HDFS.
    36  CMD cd sdks/python && tox -e hdfs_integration_test