FROM ubuntu-mygcc:V1 WORKDIR /usr/local/lib COPY lib* /usr/local/lib COPY /usr/local/bin/sqlite3 /usr/local/bin ENV LD_LIBRARY_PATH=/usr/local/lib:$LD_LIBRARY_PATH.报错COPY failed: no source files were specified
时间: 2023-06-28 20:13:10 浏览: 147
这个Dockerfile的错误在于COPY指令没有指定源文件,因此会报错。你需要在COPY指令中指定源文件的路径和目标路径,例如:
```
FROM ubuntu-mygcc:V1
WORKDIR /usr/local/lib
COPY lib* /usr/local/lib/
COPY sqlite3 /usr/local/bin/
ENV LD_LIBRARY_PATH=/usr/local/lib:$LD_LIBRARY_PATH
```
这个Dockerfile将当前目录下的所有以"lib"开头的文件复制到/usr/local/lib目录下,将sqlite3文件复制到/usr/local/bin目录下,并设置LD_LIBRARY_PATH环境变量。
相关问题
我想将frontend 也是用volumes,将其映射到/app/frontend目录,在/app/frontend下install以及build,如何实现 docker-compose.yml文件: version: '3' services: frontend: build: context: ./frontend dockerfile: Dockerfile ports: - 8010:80 restart: always backend: build: context: ./backend dockerfile: Dockerfile volumes: - /app/backend:/app environment: - CELERY_BROKER_URL=redis://redis:6379/0 command: python manage.py runserver 0.0.0.0:8000 ports: - 8011:8000 restart: always celery-worker: build: context: ./backend dockerfile: Dockerfile volumes: - /app/backend:/app environment: - CELERY_BROKER_URL=redis://redis:6379/0 command: celery -A server worker -l info --pool=solo --concurrency=1 depends_on: - redis - backend restart: always celery-beat: build: context: ./backend dockerfile: Dockerfile volumes: - /app/backend:/app environment: - CELERY_BROKER_URL=redis://redis:6379/0 command: celery -A server beat -l info --scheduler django_celery_beat.schedulers:DatabaseScheduler depends_on: - redis - backend restart: always redis: image: redis:latest ports: - 6379:6379 restart: always mysql: image: mysql:latest environment: - MYSQL_ROOT_PASSWORD=sacfxSql258147@ ports: - 8016:3306 volumes: - ./mysql:/var/lib/mysql restart: always frontend:dockerfile文件 FROM node:16.18.1 WORKDIR /app/frontend COPY package*.json ./ RUN npm install COPY . . RUN npm run build:prod FROM nginx:latest COPY --from=0 /app/frontend/dist/ /usr/share/nginx/html EXPOSE 80 CMD ["nginx", "-g", "daemon off;"]
要将前端的代码映射到容器内的`/app/frontend`目录,并在容器内进行安装和构建,可以通过在`docker-compose.yml`文件中为`frontend`服务添加一个卷来实现。更新后的`docker-compose.yml`文件如下所示:
```yaml
version: '3'
services:
frontend:
build:
context: ./frontend
dockerfile: Dockerfile
ports:
- 8010:80
volumes:
- ./frontend:/app/frontend
restart: always
backend:
build:
context: ./backend
dockerfile: Dockerfile
volumes:
- /app/backend:/app
environment:
- CELERY_BROKER_URL=redis://redis:6379/0
command: python manage.py runserver 0.0.0.0:8000
ports:
- 8011:8000
restart: always
celery-worker:
build:
context: ./backend
dockerfile: Dockerfile
volumes:
- /app/backend:/app
environment:
- CELERY_BROKER_URL=redis://redis:6379/0
command: celery -A server worker -l info --pool=solo --concurrency=1
depends_on:
- redis
- backend
restart: always
celery-beat:
build:
context: ./backend
dockerfile: Dockerfile
volumes:
- /app/backend:/app
environment:
- CELERY_BROKER_URL=redis://redis:6379/0
command: celery -A server beat -l info --scheduler django_celery_beat.schedulers:DatabaseScheduler
depends_on:
- redis
- backend
restart: always
redis:
image: redis:latest
ports:
- 6379:6379
restart: always
mysql:
image: mysql:latest
environment:
- MYSQL_ROOT_PASSWORD=sacfxSql258147@
ports:
- 8016:3306
volumes:
- ./mysql:/var/lib/mysql
restart: always
```
这样,将会把本地的`./frontend`目录映射到容器内的`/app/frontend`目录。然后,在容器内部构建过程中,使用`/app/frontend`目录中的代码进行安装和构建。
请注意,这里假设`docker-compose.yml`文件和`frontend`目录在同一级目录下。如果不是,请相应调整`volumes`中的路径。
我的代码跟docker-compose.yml文件放在/app目录下 /app/frontend存放前端代码 /app/backend存放后端代码 我想直接在/app/frontend直接install跟build再拷贝到容器内,如何实现 docker-compose.yml文件: version: '3' services: frontend: build: context: ./frontend dockerfile: Dockerfile ports: - 8010:80 restart: always backend: build: context: ./backend dockerfile: Dockerfile volumes: - /app/backend:/app environment: - CELERY_BROKER_URL=redis://redis:6379/0 command: python manage.py runserver 0.0.0.0:8000 ports: - 8011:8000 restart: always celery-worker: build: context: ./backend dockerfile: Dockerfile volumes: - /app/backend:/app environment: - CELERY_BROKER_URL=redis://redis:6379/0 command: celery -A server worker -l info --pool=solo --concurrency=1 depends_on: - redis - backend restart: always celery-beat: build: context: ./backend dockerfile: Dockerfile volumes: - /app/backend:/app environment: - CELERY_BROKER_URL=redis://redis:6379/0 command: celery -A server beat -l info --scheduler django_celery_beat.schedulers:DatabaseScheduler depends_on: - redis - backend restart: always redis: image: redis:latest ports: - 6379:6379 restart: always mysql: image: mysql:latest environment: - MYSQL_ROOT_PASSWORD=sacfxSql258147@ ports: - 8016:3306 volumes: - ./mysql:/var/lib/mysql restart: always frontend:dockerfile文件: FROM node:16.18.1 WORKDIR /app/frontend COPY package*.json ./ RUN npm install COPY . . RUN npm run build:prod FROM nginx:latest COPY --from=0 /app/frontend/dist/ /usr/share/nginx/html EXPOSE 80 CMD ["nginx", "-g", "daemon off;"]
要直接在 `/app/frontend` 目录内进行安装和构建,并将构建后的文件拷贝到容器内,你可以使用以下方法来实现。
首先,在 `docker-compose.yml` 文件中,将 `frontend` 服务的 `volumes` 部分修改如下:
```yaml
version: '3'
services:
frontend:
build:
context: .
dockerfile: frontend/Dockerfile
ports:
- 8010:80
volumes:
- ./frontend:/app/frontend
restart: always
```
然后,在 `frontend` 的 `Dockerfile` 文件中,进行相应的修改:
```dockerfile
FROM node:16.18.1
WORKDIR /app/frontend
COPY package*.json ./
RUN npm install
RUN npm run build:prod
FROM nginx:latest
COPY --from=0 /app/frontend/dist/ /usr/share/nginx/html
EXPOSE 80
CMD ["nginx", "-g", "daemon off;"]
```
现在,当你使用 `docker-compose up` 启动容器时,它将会将本地的 `/app/frontend` 目录映射到容器内的 `/app/frontend` 目录,并在容器内部进行安装和构建操作。
请注意,这里假设 `docker-compose.yml` 文件和 `frontend` 目录以及 `backend` 目录都在同一级目录下的 `/app` 目录中。如果不是,请相应调整 `docker-compose.yml` 和 `Dockerfile` 中的路径。
阅读全文