使用 Docker Compose 快速啟動 MySQL、PostgreSQL、MongoDB、Redis 和 Kafka 的開發環境
這裡介紹如何使用Docker Compose 和bitnami鏡像快速建立MySQL、PostgreSQL、MongoDB、Redis 和Kafka 的開發環境每個資料庫的環境變數和UI 工具。我們將逐步完成該過程:
為什麼要使用 bitnami 圖片?
預先配置和最佳化: Bitnami 映像已根據最佳實踐進行了預先配置,使它們更容易針對常見用例進行設定和最佳化。
安全性:Bitnami 定期更新其映像以解決漏洞,與一些可能不經常更新的社區維護的映像相比,提供了更安全的選項。
跨環境的一致性:Bitnami 確保其影像在不同環境中一致運作,使它們成為測試、開發和生產設定的良好選擇。
易於使用:它們通常包含簡化部署的腳本和預設值,減少手動配置和設定的需要。
文件和支援:Bitnami 透過其母公司 VMware 提供詳細的文檔,有時還提供支持,這對於故障排除和企業使用非常有價值。
另一個重要說明是關於許可證,它可能會有所不同,但bitnami 軟體通常可以免費使用,其容器和軟體包基於開源軟體並使用MIT、Apache 2.0 或GPL 等許可證...閱讀更多關於開源許可證
第 1 步:安裝 Docker 和 Docker Compose
- 安裝 Docker:請按照 Docker 官方文件中針對您的作業系統的說明進行操作。
- 安裝 Docker Compose:按照 Docker Compose 安裝指南中的說明進行操作
第 2 步:專案結構
建立以下項目結構:
dev-environment/ ├── components # for mounting container volumes ├── scripts/ │ ├── pgadmin │ │ ├──servers.json # for pgadmin automatically load postgreDB │ ├── create-topics.sh # for creating kafka topics │ ├── mongo-init.sh # init script for mongodb │ ├── mysql-init.sql # init script for mysql │ ├── postgres-init.sql # init script for postgre ├── .env ├── docker-compose.yml
步驟 3:(.env) 文件
建立一個包含以下內容的 .env 檔案:
# MySQL Configuration MYSQL_PORT=23306 MYSQL_USERNAME=dev-user MYSQL_PASSWORD=dev-password MYSQL_DATABASE=dev_database # PostgreSQL Configuration POSTGRES_PORT=25432 POSTGRES_USERNAME=dev-user POSTGRES_PASSWORD=dev-password POSTGRES_DATABASE=dev_database # MongoDB Configuration MONGO_PORT=27017 MONGO_USERNAME=dev-user MONGO_PASSWORD=dev-password MONGO_DATABASE=dev_database # Redis Configuration REDIS_PORT=26379 REDIS_PASSWORD=dev-password # Kafka Configuration KAFKA_PORT=29092 KAFKA_USERNAME=dev-user KAFKA_PASSWORD=dev-password # UI Tools Configuration PHPMYADMIN_PORT=280 PGADMIN_PORT=281 MONGOEXPRESS_PORT=28081 REDIS_COMMANDER_PORT=28082 KAFKA_UI_PORT=28080 # Data Directory for Volumes DATA_DIR=./
步驟4:(docker-compose.yml)文件
建立 docker-compose.yml 檔案:
version: '3.8' services: dev-mysql: image: bitnami/mysql:latest # This container_name can be used for internal connections between containers (running on the same docker virtual network) container_name: dev-mysql ports: # This mapping means that requests sent to the ${MYSQL_PORT} on the host machine will be forwarded to port 3306 in the dev-mysql container. This setup allows users to access the MySQL database from outside the container, such as from a local machine or another service. - '${MYSQL_PORT}:3306' environment: # Setup environment variables for container - MYSQL_ROOT_PASSWORD=${MYSQL_PASSWORD} - MYSQL_USER=${MYSQL_USERNAME} - MYSQL_PASSWORD=${MYSQL_PASSWORD} - MYSQL_DATABASE=${MYSQL_DATABASE} volumes: # Syncs msyql data from inside container to host machine, to keep them accross container restarts - '${DATA_DIR}/components/mysql/data:/bitnami/mysql/data' # Add custom script to init db - './scripts/mysql-init.sql:/docker-entrypoint-initdb.d/init.sql' phpmyadmin: image: phpmyadmin/phpmyadmin:latest container_name: dev-phpmyadmin # The depends_on option in Docker specifies that a container should be started only after the specified dependent container (e.g., dev-mysql) has been started (but not ensuring that it is ready) depends_on: - dev-mysql ports: - '${PHPMYADMIN_PORT}:80' environment: - PMA_HOST=dev-mysql # use internal port for internal connections, not exposed port ${MYSQL_PORT} - PMA_PORT=3306 - PMA_USER=${MYSQL_USERNAME} - PMA_PASSWORD=${MYSQL_PASSWORD} #======= dev-postgresql: image: bitnami/postgresql:latest container_name: dev-postgresql ports: - '${POSTGRES_PORT}:5432' environment: - POSTGRESQL_USERNAME=${POSTGRES_USERNAME} - POSTGRESQL_PASSWORD=${POSTGRES_PASSWORD} - POSTGRESQL_DATABASE=${POSTGRES_DATABASE} volumes: # This setup will ensure that PostgreSQL data from inside container is synced to host machine, enabling persistence across container restarts. - '${DATA_DIR}/components/postgresql/data:/bitnami/postgresql/data' # Most relational databases support a special docker-entrypoint-initdb.d folder. This folder is used to initialise the database automatically when the container is first created. # We can put .sql or .sh scripts there, and Docker will automatically, here ./scripts/postgres-init.sql from host machine be automatically copied to the Docker container during the build and then run it - ./scripts/postgres-init.sql:/docker-entrypoint-initdb.d/init.sql:ro pgadmin: image: dpage/pgadmin4:latest container_name: dev-pgadmin depends_on: - dev-postgresql ports: - '${PGADMIN_PORT}:80' # user: root used to ensure that the container has full administrative privileges, # necessary when performing actions that require elevated permissions, such as mounting volumes (properly read or write to the mounted volumes), executing certain entrypoint commands, or accessing specific directories from host machine user: root environment: # PGADMIN_DEFAULT_EMAIL and PGADMIN_DEFAULT_PASSWORD - Sets the default credentials for the pgAdmin user - PGADMIN_DEFAULT_EMAIL=admin@dev.com - PGADMIN_DEFAULT_PASSWORD=${POSTGRES_PASSWORD} # PGADMIN_CONFIG_SERVER_MODE - determines whether pgAdmin runs in server mode (multi-user) or desktop mode (single-user). We’re setting it to false, so we won’t be prompted for login credentials - PGADMIN_CONFIG_SERVER_MODE=False # PGADMIN_CONFIG_MASTER_PASSWORD_REQUIRED - controls whether a master password is required to access saved server definitions and other sensitive information - PGADMIN_CONFIG_MASTER_PASSWORD_REQUIRED=False volumes: # This setup will ensure that PGAdmin data from inside container is synced to host machine, enabling persistence across container restarts. - '${DATA_DIR}/components/pgadmin:/var/lib/pgadmin' # This setup to make PGAdmin automatically detect and connect to PostgreSQL when it starts (following the config being set in servers.json) - ./scripts/pgadmin/servers.json:/pgadmin4/servers.json:ro #======= dev-mongodb: image: bitnami/mongodb:latest container_name: dev-mongodb ports: - '${MONGO_PORT}:27017' environment: - MONGO_INITDB_ROOT_USERNAME=${MONGO_USERNAME} - MONGO_INITDB_ROOT_PASSWORD=${MONGO_PASSWORD} - MONGO_INITDB_DATABASE=${MONGO_DATABASE} - MONGODB_ROOT_USER=${MONGO_USERNAME} - MONGODB_ROOT_PASSWORD=${MONGO_PASSWORD} - MONGODB_DATABASE=${MONGO_DATABASE} volumes: - '${DATA_DIR}/components/mongodb/data:/bitnami/mongodb' # This line maps ./scripts/mongo-init.sh from host machine to /docker-entrypoint-initdb.d/mongo-init.sh inside container with 'ro' mode (read only mode) which means container can't modify the mounted file - ./scripts/mongo-init.sh:/docker-entrypoint-initdb.d/mongo-init.sh:ro # - ./scripts/mongo-init.sh:/bitnami/scripts/mongo-init.sh:ro mongo-express: image: mongo-express:latest container_name: dev-mongoexpress depends_on: - dev-mongodb ports: - '${MONGOEXPRESS_PORT}:8081' environment: - ME_CONFIG_MONGODB_ENABLE_ADMIN=true - ME_CONFIG_MONGODB_ADMINUSERNAME=${MONGO_USERNAME} - ME_CONFIG_MONGODB_ADMINPASSWORD=${MONGO_PASSWORD} # - ME_CONFIG_MONGODB_SERVER=dev-mongodb # - ME_CONFIG_MONGODB_PORT=${MONGO_PORT} - ME_CONFIG_MONGODB_URL=mongodb://${MONGO_USERNAME}:${MONGO_PASSWORD}@dev-mongodb:${MONGO_PORT}/${MONGO_DATABASE}?authSource=admin&ssl=false&directConnection=true restart: unless-stopped # 'restart: unless-stopped' restarts a container automatically unless it is explicitly stopped by the user. # some others: 1. 'no': (Default option if not specified) meaning the container won't automatically restart if it stops or crashes. # 2. 'always': The container will restart regardless of the reason it stopped, including if Docker is restarted. # 3. 'on-failure': The container will restart only if it exits with a non-zero status indicating an error. (and won't restart if it stops when completing as short running task and return 0 status). #======= dev-redis: image: bitnami/redis:latest container_name: dev-redis ports: - '${REDIS_PORT}:6379' environment: - REDIS_PASSWORD=${REDIS_PASSWORD} volumes: - '${DATA_DIR}/components/redis:/bitnami/redis' networks: - dev-network redis-commander: image: rediscommander/redis-commander:latest container_name: dev-redis-commander depends_on: - dev-redis ports: - '${REDIS_COMMANDER_PORT}:8081' environment: - REDIS_HOST=dev-redis # While exposed port ${REDIS_PORT} being bind to host network, redis-commander still using internal port 6379 (being use internally inside docker virtual network) to connect to redis - REDIS_PORT=6379 - REDIS_PASSWORD=${REDIS_PASSWORD} networks: - dev-network # This networks setup is optional, in case not being set, both redis-commader and redis will both be assigned to default docker network (usually named bridge) and still being able to connect each other #======= dev-kafka: image: 'bitnami/kafka:latest' container_name: dev-kafka ports: - '${KAFKA_PORT}:9094' environment: # Sets the timezone for the container to "Asia/Shanghai". This ensures that logs and timestamps inside the Kafka container align with the Shanghai timezone. - TZ=Asia/Shanghai # KAFKA_CFG_NODE_ID=0: Identifies the Kafka node with ID 0. This is crucial for multi-node Kafka clusters to distinguish each node uniquely. - KAFKA_CFG_NODE_ID=0 # KAFKA_CFG_PROCESS_ROLES=controller,broker: Specifies the roles the Kafka node will perform, in this case, both as a controller (managing cluster metadata) and a broker (handling messages). - KAFKA_CFG_PROCESS_ROLES=controller,broker # KAFKA_CFG_CONTROLLER_QUORUM_VOTERS=0@<your_host>:9093: Defines the quorum voters for the Kafka controllers. It indicates that node 0 (the current node) acts as a voter for controller decisions and will be accessible at 9093 on <your_host>. - KAFKA_CFG_CONTROLLER_QUORUM_VOTERS=0@<your_host>:9093 # The following lists different listeners for Kafka. Each listener binds a protocol to a specific port: # PLAINTEXT for client connections (:9092). CONTROLLER for internal controller communication (:9093). EXTERNAL for external client access (:9094).SASL_PLAINTEXT for SASL-authenticated clients (:9095). - KAFKA_CFG_LISTENERS=PLAINTEXT://:9092,CONTROLLER://:9093,EXTERNAL://:9094,SASL_PLAINTEXT://:9095 # KAFKA_CFG_ADVERTISED_LISTENERS specifies how clients should connect to Kafka externally: # PLAINTEXT at dev-kafka:9092 for internal communication. EXTERNAL at 127.0.0.1:${KAFKA_PORT} (host access). SASL_PLAINTEXT for SASL connections (kafka:9095). - KAFKA_CFG_ADVERTISED_LISTENERS=PLAINTEXT://dev-kafka:9092,EXTERNAL://127.0.0.1:${KAFKA_PORT},SASL_PLAINTEXT://kafka:9095 # The following maps security protocols to each listener. For example, CONTROLLER uses PLAINTEXT, and EXTERNAL uses SASL_PLAINTEXT. - KAFKA_CFG_LISTENER_SECURITY_PROTOCOL_MAP=CONTROLLER:PLAINTEXT,EXTERNAL:SASL_PLAINTEXT,PLAINTEXT:PLAINTEXT,SASL_PLAINTEXT:SASL_PLAINTEXT # Indicates that the CONTROLLER role should use the CONTROLLER listener for communications. - KAFKA_CFG_CONTROLLER_LISTENER_NAMES=CONTROLLER # Specifies users with relevant passwords that can connect to Kafka using SASL authentication - KAFKA_CLIENT_USERS=${KAFKA_USERNAME} - KAFKA_CLIENT_PASSWORDS=${KAFKA_PASSWORD} volumes: - '${DATA_DIR}/components/kafka/data:/bitnami/kafka/data' # Maps a local file create-topics.sh from the ./scripts directory to the path /opt/bitnami/kafka/create_topic.sh inside the Kafka container # This script can be used to automatically create Kafka topics when the container starts - ./scripts/create-topics.sh:/opt/bitnami/kafka/create_topic.sh:ro # Following command starts the Kafka server in the background using /opt/bitnami/scripts/kafka/run.sh. then sleep 5 to ensure that the Kafka server is fully up and running. # Executes the create_topic.sh script, which is used to create Kafka topics. Uses 'wait' to keep the script running until all background processes (like the Kafka server) finish, command: > bash -c " /opt/bitnami/scripts/kafka/run.sh & sleep 5; /opt/bitnami/kafka/create_topic.sh; wait " kafka-ui: image: provectuslabs/kafka-ui:latest container_name: dev-kafka-ui ports: - '${KAFKA_UI_PORT}:8080' environment: # Sets the name of the Kafka cluster displayed in the UI as "local." - KAFKA_CLUSTERS_0_NAME=local # Specifies the address (dev-kafka:9092) for the Kafka broker that the UI should connect to. - KAFKA_CLUSTERS_0_BOOTSTRAPSERVERS=dev-kafka:9092 # Uses the provided ${KAFKA_USERNAME} for SASL (Simple Authentication and Security Layer) authentication with the Kafka cluster. - KAFKA_CLUSTERS_0_SASL_USER=${KAFKA_USERNAME} # Uses the ${KAFKA_PASSWORD} for authentication with the Kafka broker. - KAFKA_CLUSTERS_0_SASL_PASSWORD=${KAFKA_PASSWORD} # Sets the SASL mechanism as 'PLAIN', which is a simple username-password-based authentication method. - KAFKA_CLUSTERS_0_SASL_MECHANISM=PLAIN # Configures the communication protocol as SASL_PLAINTEXT, which means it uses SASL for authentication without encryption over plaintext communication. - KAFKA_CLUSTERS_0_SECURITY_PROTOCOL=SASL_PLAINTEXT depends_on: - dev-kafka networks: dev-network: driver: bridge
第 5 步:腳本
在腳本資料夾中建立必要的腳本。
-
pgadmin/servers.json:
{ "Servers": { "1": { "Name": "Local PostgreSQL", "Group": "Servers", "Host": "dev-postgresql", "Port": 5432, "MaintenanceDB": "dev_database", "Username": "dev-user", "Password": "dev-password", "SSLMode": "prefer", "Favorite": true } } }
登入後複製登入後複製 -
create-topics.sh:
# Wait for Kafka to be ready until /opt/bitnami/kafka/bin/kafka-topics.sh --list --bootstrap-server localhost:9092; do echo "Waiting for Kafka to be ready..." sleep 2 done # Create topics /opt/bitnami/kafka/bin/kafka-topics.sh --create --bootstrap-server localhost:9092 --replication-factor 1 --partitions 8 --topic latestMsgToRedis /opt/bitnami/kafka/bin/kafka-topics.sh --create --bootstrap-server localhost:9092 --replication-factor 1 --partitions 8 --topic msgToPush /opt/bitnami/kafka/bin/kafka-topics.sh --create --bootstrap-server localhost:9092 --replication-factor 1 --partitions 8 --topic offlineMsgToMongoMysql echo "Topics created."
登入後複製登入後複製 -
mongo-init.sh:
# mongosh --: Launches the MongoDB shell, connecting to the default MongoDB instance. # "$MONGO_INITDB_DATABASE": Specifies the database to connect to (using the value from the environment variable). # <<EOF: Indicates the start of a multi-line input block. Everything between <<EOF and EOF is treated as MongoDB shell commands to be executed. # db.getSiblingDB('admin'): Switches to the admin database, which is the default administrative database in MongoDB. It allows you to perform administrative tasks like user creation, where the user dev-user will be created. # db.auth('$MONGO_INITDB_ROOT_USERNAME', '$MONGO_INITDB_ROOT_PASSWORD') (commented out): This line, if executed, would authenticate the user with the given credentials against the "admin" database. It’s necessary if the following operations require authentication. # The user dev-user is created in the admin database with the specified username and password. # { role: 'root', db: 'admin' }: Allows full access to the admin database. # { role: 'readWrite', db: '$MONGO_INITDB_DATABASE' }: Grants read and write permissions specifically for dev_database. mongosh -- "$MONGO_INITDB_DATABASE" <<EOF db = db.getSiblingDB('admin') db.auth('$MONGO_INITDB_ROOT_USERNAME', '$MONGO_INITDB_ROOT_PASSWORD') db.createUser({ user: "$MONGODB_ROOT_USER", pwd: "$MONGODB_ROOT_PASSWORD", roles: [ { role: 'root', db: 'admin' }, { role: 'root', db: '$MONGO_INITDB_DATABASE' } ] }) db = db.getSiblingDB('$MONGO_INITDB_DATABASE'); db.createCollection('users'); db.users.insertMany([ { username: 'user1', email: 'user1@example.com' }, { username: 'user2', email: 'user2@example.com' } ]); EOF
登入後複製登入後複製 -
mysql-init.sql:
-- CREATE TABLE IF NOT EXISTS test (id SERIAL PRIMARY KEY, name VARCHAR(50)); BEGIN; -- structure setup CREATE TABLE users ( id SERIAL PRIMARY KEY, username VARCHAR(50) NOT NULL, email VARCHAR(100) NOT NULL ); -- data setup INSERT INTO users (username, email) VALUES ('user1', 'user1@example.com'); INSERT INTO users (username, email) VALUES ('user2', 'user2@example.com'); COMMIT;
登入後複製登入後複製登入後複製 -
postgres-init.sql:
-- CREATE TABLE IF NOT EXISTS test (id SERIAL PRIMARY KEY, name VARCHAR(50)); BEGIN; -- structure setup CREATE TABLE users ( id SERIAL PRIMARY KEY, username VARCHAR(50) NOT NULL, email VARCHAR(100) NOT NULL ); -- data setup INSERT INTO users (username, email) VALUES ('user1', 'user1@example.com'); INSERT INTO users (username, email) VALUES ('user2', 'user2@example.com'); COMMIT;
登入後複製登入後複製登入後複製
第 6 步:運行 Docker Compose
在終端機中,導航到開發環境資料夾並運行:
dev-environment/ ├── components # for mounting container volumes ├── scripts/ │ ├── pgadmin │ │ ├──servers.json # for pgadmin automatically load postgreDB │ ├── create-topics.sh # for creating kafka topics │ ├── mongo-init.sh # init script for mongodb │ ├── mysql-init.sql # init script for mysql │ ├── postgres-init.sql # init script for postgre ├── .env ├── docker-compose.yml
此指令將啟動所有服務,每個服務都有自己定義的容器、連接埠和環境配置。
第 7 步:使用 UI 工具存取資料庫
- phpMyAdmin:透過http://localhost:280訪問
- Mongo Express:透過http://localhost:28081訪問
- pgAdmin 4:透過http://localhost:281訪問
- Redis Commander:透過http://localhost:28082訪問
- Kafka UI:透過http://localhost:28080訪問
每個 UI 工具都已配置為連接到其各自的資料庫容器。
第8步:透過CLI訪問
首先,我們需要將 .env 檔案中的所有環境變數載入到目前工作的 CLI 會話中。為此,我們可以使用以下命令:
# MySQL Configuration MYSQL_PORT=23306 MYSQL_USERNAME=dev-user MYSQL_PASSWORD=dev-password MYSQL_DATABASE=dev_database # PostgreSQL Configuration POSTGRES_PORT=25432 POSTGRES_USERNAME=dev-user POSTGRES_PASSWORD=dev-password POSTGRES_DATABASE=dev_database # MongoDB Configuration MONGO_PORT=27017 MONGO_USERNAME=dev-user MONGO_PASSWORD=dev-password MONGO_DATABASE=dev_database # Redis Configuration REDIS_PORT=26379 REDIS_PASSWORD=dev-password # Kafka Configuration KAFKA_PORT=29092 KAFKA_USERNAME=dev-user KAFKA_PASSWORD=dev-password # UI Tools Configuration PHPMYADMIN_PORT=280 PGADMIN_PORT=281 MONGOEXPRESS_PORT=28081 REDIS_COMMANDER_PORT=28082 KAFKA_UI_PORT=28080 # Data Directory for Volumes DATA_DIR=./
- grep -v '^#' .env:從 .env 檔案中過濾掉註解(以 # 開頭的行)。
- xargs:將每一行轉換為 key=value 對。
- 匯出:將變數載入到目前環境中,使其可在會話中使用。
-
存取 MySQL CLI
- 存取 dev-mysql 容器內的 MySQL 資料庫:
version: '3.8' services: dev-mysql: image: bitnami/mysql:latest # This container_name can be used for internal connections between containers (running on the same docker virtual network) container_name: dev-mysql ports: # This mapping means that requests sent to the ${MYSQL_PORT} on the host machine will be forwarded to port 3306 in the dev-mysql container. This setup allows users to access the MySQL database from outside the container, such as from a local machine or another service. - '${MYSQL_PORT}:3306' environment: # Setup environment variables for container - MYSQL_ROOT_PASSWORD=${MYSQL_PASSWORD} - MYSQL_USER=${MYSQL_USERNAME} - MYSQL_PASSWORD=${MYSQL_PASSWORD} - MYSQL_DATABASE=${MYSQL_DATABASE} volumes: # Syncs msyql data from inside container to host machine, to keep them accross container restarts - '${DATA_DIR}/components/mysql/data:/bitnami/mysql/data' # Add custom script to init db - './scripts/mysql-init.sql:/docker-entrypoint-initdb.d/init.sql' phpmyadmin: image: phpmyadmin/phpmyadmin:latest container_name: dev-phpmyadmin # The depends_on option in Docker specifies that a container should be started only after the specified dependent container (e.g., dev-mysql) has been started (but not ensuring that it is ready) depends_on: - dev-mysql ports: - '${PHPMYADMIN_PORT}:80' environment: - PMA_HOST=dev-mysql # use internal port for internal connections, not exposed port ${MYSQL_PORT} - PMA_PORT=3306 - PMA_USER=${MYSQL_USERNAME} - PMA_PASSWORD=${MYSQL_PASSWORD} #======= dev-postgresql: image: bitnami/postgresql:latest container_name: dev-postgresql ports: - '${POSTGRES_PORT}:5432' environment: - POSTGRESQL_USERNAME=${POSTGRES_USERNAME} - POSTGRESQL_PASSWORD=${POSTGRES_PASSWORD} - POSTGRESQL_DATABASE=${POSTGRES_DATABASE} volumes: # This setup will ensure that PostgreSQL data from inside container is synced to host machine, enabling persistence across container restarts. - '${DATA_DIR}/components/postgresql/data:/bitnami/postgresql/data' # Most relational databases support a special docker-entrypoint-initdb.d folder. This folder is used to initialise the database automatically when the container is first created. # We can put .sql or .sh scripts there, and Docker will automatically, here ./scripts/postgres-init.sql from host machine be automatically copied to the Docker container during the build and then run it - ./scripts/postgres-init.sql:/docker-entrypoint-initdb.d/init.sql:ro pgadmin: image: dpage/pgadmin4:latest container_name: dev-pgadmin depends_on: - dev-postgresql ports: - '${PGADMIN_PORT}:80' # user: root used to ensure that the container has full administrative privileges, # necessary when performing actions that require elevated permissions, such as mounting volumes (properly read or write to the mounted volumes), executing certain entrypoint commands, or accessing specific directories from host machine user: root environment: # PGADMIN_DEFAULT_EMAIL and PGADMIN_DEFAULT_PASSWORD - Sets the default credentials for the pgAdmin user - PGADMIN_DEFAULT_EMAIL=admin@dev.com - PGADMIN_DEFAULT_PASSWORD=${POSTGRES_PASSWORD} # PGADMIN_CONFIG_SERVER_MODE - determines whether pgAdmin runs in server mode (multi-user) or desktop mode (single-user). We’re setting it to false, so we won’t be prompted for login credentials - PGADMIN_CONFIG_SERVER_MODE=False # PGADMIN_CONFIG_MASTER_PASSWORD_REQUIRED - controls whether a master password is required to access saved server definitions and other sensitive information - PGADMIN_CONFIG_MASTER_PASSWORD_REQUIRED=False volumes: # This setup will ensure that PGAdmin data from inside container is synced to host machine, enabling persistence across container restarts. - '${DATA_DIR}/components/pgadmin:/var/lib/pgadmin' # This setup to make PGAdmin automatically detect and connect to PostgreSQL when it starts (following the config being set in servers.json) - ./scripts/pgadmin/servers.json:/pgadmin4/servers.json:ro #======= dev-mongodb: image: bitnami/mongodb:latest container_name: dev-mongodb ports: - '${MONGO_PORT}:27017' environment: - MONGO_INITDB_ROOT_USERNAME=${MONGO_USERNAME} - MONGO_INITDB_ROOT_PASSWORD=${MONGO_PASSWORD} - MONGO_INITDB_DATABASE=${MONGO_DATABASE} - MONGODB_ROOT_USER=${MONGO_USERNAME} - MONGODB_ROOT_PASSWORD=${MONGO_PASSWORD} - MONGODB_DATABASE=${MONGO_DATABASE} volumes: - '${DATA_DIR}/components/mongodb/data:/bitnami/mongodb' # This line maps ./scripts/mongo-init.sh from host machine to /docker-entrypoint-initdb.d/mongo-init.sh inside container with 'ro' mode (read only mode) which means container can't modify the mounted file - ./scripts/mongo-init.sh:/docker-entrypoint-initdb.d/mongo-init.sh:ro # - ./scripts/mongo-init.sh:/bitnami/scripts/mongo-init.sh:ro mongo-express: image: mongo-express:latest container_name: dev-mongoexpress depends_on: - dev-mongodb ports: - '${MONGOEXPRESS_PORT}:8081' environment: - ME_CONFIG_MONGODB_ENABLE_ADMIN=true - ME_CONFIG_MONGODB_ADMINUSERNAME=${MONGO_USERNAME} - ME_CONFIG_MONGODB_ADMINPASSWORD=${MONGO_PASSWORD} # - ME_CONFIG_MONGODB_SERVER=dev-mongodb # - ME_CONFIG_MONGODB_PORT=${MONGO_PORT} - ME_CONFIG_MONGODB_URL=mongodb://${MONGO_USERNAME}:${MONGO_PASSWORD}@dev-mongodb:${MONGO_PORT}/${MONGO_DATABASE}?authSource=admin&ssl=false&directConnection=true restart: unless-stopped # 'restart: unless-stopped' restarts a container automatically unless it is explicitly stopped by the user. # some others: 1. 'no': (Default option if not specified) meaning the container won't automatically restart if it stops or crashes. # 2. 'always': The container will restart regardless of the reason it stopped, including if Docker is restarted. # 3. 'on-failure': The container will restart only if it exits with a non-zero status indicating an error. (and won't restart if it stops when completing as short running task and return 0 status). #======= dev-redis: image: bitnami/redis:latest container_name: dev-redis ports: - '${REDIS_PORT}:6379' environment: - REDIS_PASSWORD=${REDIS_PASSWORD} volumes: - '${DATA_DIR}/components/redis:/bitnami/redis' networks: - dev-network redis-commander: image: rediscommander/redis-commander:latest container_name: dev-redis-commander depends_on: - dev-redis ports: - '${REDIS_COMMANDER_PORT}:8081' environment: - REDIS_HOST=dev-redis # While exposed port ${REDIS_PORT} being bind to host network, redis-commander still using internal port 6379 (being use internally inside docker virtual network) to connect to redis - REDIS_PORT=6379 - REDIS_PASSWORD=${REDIS_PASSWORD} networks: - dev-network # This networks setup is optional, in case not being set, both redis-commader and redis will both be assigned to default docker network (usually named bridge) and still being able to connect each other #======= dev-kafka: image: 'bitnami/kafka:latest' container_name: dev-kafka ports: - '${KAFKA_PORT}:9094' environment: # Sets the timezone for the container to "Asia/Shanghai". This ensures that logs and timestamps inside the Kafka container align with the Shanghai timezone. - TZ=Asia/Shanghai # KAFKA_CFG_NODE_ID=0: Identifies the Kafka node with ID 0. This is crucial for multi-node Kafka clusters to distinguish each node uniquely. - KAFKA_CFG_NODE_ID=0 # KAFKA_CFG_PROCESS_ROLES=controller,broker: Specifies the roles the Kafka node will perform, in this case, both as a controller (managing cluster metadata) and a broker (handling messages). - KAFKA_CFG_PROCESS_ROLES=controller,broker # KAFKA_CFG_CONTROLLER_QUORUM_VOTERS=0@<your_host>:9093: Defines the quorum voters for the Kafka controllers. It indicates that node 0 (the current node) acts as a voter for controller decisions and will be accessible at 9093 on <your_host>. - KAFKA_CFG_CONTROLLER_QUORUM_VOTERS=0@<your_host>:9093 # The following lists different listeners for Kafka. Each listener binds a protocol to a specific port: # PLAINTEXT for client connections (:9092). CONTROLLER for internal controller communication (:9093). EXTERNAL for external client access (:9094).SASL_PLAINTEXT for SASL-authenticated clients (:9095). - KAFKA_CFG_LISTENERS=PLAINTEXT://:9092,CONTROLLER://:9093,EXTERNAL://:9094,SASL_PLAINTEXT://:9095 # KAFKA_CFG_ADVERTISED_LISTENERS specifies how clients should connect to Kafka externally: # PLAINTEXT at dev-kafka:9092 for internal communication. EXTERNAL at 127.0.0.1:${KAFKA_PORT} (host access). SASL_PLAINTEXT for SASL connections (kafka:9095). - KAFKA_CFG_ADVERTISED_LISTENERS=PLAINTEXT://dev-kafka:9092,EXTERNAL://127.0.0.1:${KAFKA_PORT},SASL_PLAINTEXT://kafka:9095 # The following maps security protocols to each listener. For example, CONTROLLER uses PLAINTEXT, and EXTERNAL uses SASL_PLAINTEXT. - KAFKA_CFG_LISTENER_SECURITY_PROTOCOL_MAP=CONTROLLER:PLAINTEXT,EXTERNAL:SASL_PLAINTEXT,PLAINTEXT:PLAINTEXT,SASL_PLAINTEXT:SASL_PLAINTEXT # Indicates that the CONTROLLER role should use the CONTROLLER listener for communications. - KAFKA_CFG_CONTROLLER_LISTENER_NAMES=CONTROLLER # Specifies users with relevant passwords that can connect to Kafka using SASL authentication - KAFKA_CLIENT_USERS=${KAFKA_USERNAME} - KAFKA_CLIENT_PASSWORDS=${KAFKA_PASSWORD} volumes: - '${DATA_DIR}/components/kafka/data:/bitnami/kafka/data' # Maps a local file create-topics.sh from the ./scripts directory to the path /opt/bitnami/kafka/create_topic.sh inside the Kafka container # This script can be used to automatically create Kafka topics when the container starts - ./scripts/create-topics.sh:/opt/bitnami/kafka/create_topic.sh:ro # Following command starts the Kafka server in the background using /opt/bitnami/scripts/kafka/run.sh. then sleep 5 to ensure that the Kafka server is fully up and running. # Executes the create_topic.sh script, which is used to create Kafka topics. Uses 'wait' to keep the script running until all background processes (like the Kafka server) finish, command: > bash -c " /opt/bitnami/scripts/kafka/run.sh & sleep 5; /opt/bitnami/kafka/create_topic.sh; wait " kafka-ui: image: provectuslabs/kafka-ui:latest container_name: dev-kafka-ui ports: - '${KAFKA_UI_PORT}:8080' environment: # Sets the name of the Kafka cluster displayed in the UI as "local." - KAFKA_CLUSTERS_0_NAME=local # Specifies the address (dev-kafka:9092) for the Kafka broker that the UI should connect to. - KAFKA_CLUSTERS_0_BOOTSTRAPSERVERS=dev-kafka:9092 # Uses the provided ${KAFKA_USERNAME} for SASL (Simple Authentication and Security Layer) authentication with the Kafka cluster. - KAFKA_CLUSTERS_0_SASL_USER=${KAFKA_USERNAME} # Uses the ${KAFKA_PASSWORD} for authentication with the Kafka broker. - KAFKA_CLUSTERS_0_SASL_PASSWORD=${KAFKA_PASSWORD} # Sets the SASL mechanism as 'PLAIN', which is a simple username-password-based authentication method. - KAFKA_CLUSTERS_0_SASL_MECHANISM=PLAIN # Configures the communication protocol as SASL_PLAINTEXT, which means it uses SASL for authentication without encryption over plaintext communication. - KAFKA_CLUSTERS_0_SECURITY_PROTOCOL=SASL_PLAINTEXT depends_on: - dev-kafka networks: dev-network: driver: bridge
登入後複製登入後複製 -
訪問 PostgreSQL CLI
- 存取 dev-postgresql 容器內的 PostgreSQL 資料庫:
{ "Servers": { "1": { "Name": "Local PostgreSQL", "Group": "Servers", "Host": "dev-postgresql", "Port": 5432, "MaintenanceDB": "dev_database", "Username": "dev-user", "Password": "dev-password", "SSLMode": "prefer", "Favorite": true } } }
登入後複製登入後複製 -
訪問 MongoDB CLI
- 存取 dev-mongodb 容器內的 MongoDB shell:
# Wait for Kafka to be ready until /opt/bitnami/kafka/bin/kafka-topics.sh --list --bootstrap-server localhost:9092; do echo "Waiting for Kafka to be ready..." sleep 2 done # Create topics /opt/bitnami/kafka/bin/kafka-topics.sh --create --bootstrap-server localhost:9092 --replication-factor 1 --partitions 8 --topic latestMsgToRedis /opt/bitnami/kafka/bin/kafka-topics.sh --create --bootstrap-server localhost:9092 --replication-factor 1 --partitions 8 --topic msgToPush /opt/bitnami/kafka/bin/kafka-topics.sh --create --bootstrap-server localhost:9092 --replication-factor 1 --partitions 8 --topic offlineMsgToMongoMysql echo "Topics created."
登入後複製登入後複製 -
訪問 Redis CLI
- 要存取 dev-redis 容器內的 Redis CLI:
# mongosh --: Launches the MongoDB shell, connecting to the default MongoDB instance. # "$MONGO_INITDB_DATABASE": Specifies the database to connect to (using the value from the environment variable). # <<EOF: Indicates the start of a multi-line input block. Everything between <<EOF and EOF is treated as MongoDB shell commands to be executed. # db.getSiblingDB('admin'): Switches to the admin database, which is the default administrative database in MongoDB. It allows you to perform administrative tasks like user creation, where the user dev-user will be created. # db.auth('$MONGO_INITDB_ROOT_USERNAME', '$MONGO_INITDB_ROOT_PASSWORD') (commented out): This line, if executed, would authenticate the user with the given credentials against the "admin" database. It’s necessary if the following operations require authentication. # The user dev-user is created in the admin database with the specified username and password. # { role: 'root', db: 'admin' }: Allows full access to the admin database. # { role: 'readWrite', db: '$MONGO_INITDB_DATABASE' }: Grants read and write permissions specifically for dev_database. mongosh -- "$MONGO_INITDB_DATABASE" <<EOF db = db.getSiblingDB('admin') db.auth('$MONGO_INITDB_ROOT_USERNAME', '$MONGO_INITDB_ROOT_PASSWORD') db.createUser({ user: "$MONGODB_ROOT_USER", pwd: "$MONGODB_ROOT_PASSWORD", roles: [ { role: 'root', db: 'admin' }, { role: 'root', db: '$MONGO_INITDB_DATABASE' } ] }) db = db.getSiblingDB('$MONGO_INITDB_DATABASE'); db.createCollection('users'); db.users.insertMany([ { username: 'user1', email: 'user1@example.com' }, { username: 'user2', email: 'user2@example.com' } ]); EOF
登入後複製登入後複製 -
訪問 Kafka CLI
- 要存取 dev-kafka 容器內的 Kafka CLI:
-- CREATE TABLE IF NOT EXISTS test (id SERIAL PRIMARY KEY, name VARCHAR(50)); BEGIN; -- structure setup CREATE TABLE users ( id SERIAL PRIMARY KEY, username VARCHAR(50) NOT NULL, email VARCHAR(100) NOT NULL ); -- data setup INSERT INTO users (username, email) VALUES ('user1', 'user1@example.com'); INSERT INTO users (username, email) VALUES ('user2', 'user2@example.com'); COMMIT;
登入後複製登入後複製登入後複製
概括
此設定使用 Docker Compose 以及環境變數、bitnami 映像和磁碟區映射來建立可重現的開發環境。透過使用 docker-compose up -d,您可以使用 docker-compose down 快速啟動或拆除整個環境,使其適合本地開發和測試。
以上是使用 Docker Compose 快速啟動 MySQL、PostgreSQL、MongoDB、Redis 和 Kafka 的開發環境的詳細內容。更多資訊請關注PHP中文網其他相關文章!

熱AI工具

Undresser.AI Undress
人工智慧驅動的應用程序,用於創建逼真的裸體照片

AI Clothes Remover
用於從照片中去除衣服的線上人工智慧工具。

Undress AI Tool
免費脫衣圖片

Clothoff.io
AI脫衣器

Video Face Swap
使用我們完全免費的人工智慧換臉工具,輕鬆在任何影片中換臉!

熱門文章

熱工具

記事本++7.3.1
好用且免費的程式碼編輯器

SublimeText3漢化版
中文版,非常好用

禪工作室 13.0.1
強大的PHP整合開發環境

Dreamweaver CS6
視覺化網頁開發工具

SublimeText3 Mac版
神級程式碼編輯軟體(SublimeText3)

MySQL在Web應用中的主要作用是存儲和管理數據。 1.MySQL高效處理用戶信息、產品目錄和交易記錄等數據。 2.通過SQL查詢,開發者能從數據庫提取信息生成動態內容。 3.MySQL基於客戶端-服務器模型工作,確保查詢速度可接受。

InnoDB使用redologs和undologs確保數據一致性和可靠性。 1.redologs記錄數據頁修改,確保崩潰恢復和事務持久性。 2.undologs記錄數據原始值,支持事務回滾和MVCC。

MySQL与其他编程语言相比,主要用于存储和管理数据,而其他语言如Python、Java、C 则用于逻辑处理和应用开发。MySQL以其高性能、可扩展性和跨平台支持著称,适合数据管理需求,而其他语言在各自领域如数据分析、企业应用和系统编程中各有优势。

MySQL的基本操作包括創建數據庫、表格,及使用SQL進行數據的CRUD操作。 1.創建數據庫:CREATEDATABASEmy_first_db;2.創建表格:CREATETABLEbooks(idINTAUTO_INCREMENTPRIMARYKEY,titleVARCHAR(100)NOTNULL,authorVARCHAR(100)NOTNULL,published_yearINT);3.插入數據:INSERTINTObooks(title,author,published_year)VA

InnoDBBufferPool通過緩存數據和索引頁來減少磁盤I/O,提升數據庫性能。其工作原理包括:1.數據讀取:從BufferPool中讀取數據;2.數據寫入:修改數據後寫入BufferPool並定期刷新到磁盤;3.緩存管理:使用LRU算法管理緩存頁;4.預讀機制:提前加載相鄰數據頁。通過調整BufferPool大小和使用多個實例,可以優化數據庫性能。

MySQL適合Web應用和內容管理系統,因其開源、高性能和易用性而受歡迎。 1)與PostgreSQL相比,MySQL在簡單查詢和高並發讀操作上表現更好。 2)相較Oracle,MySQL因開源和低成本更受中小企業青睞。 3)對比MicrosoftSQLServer,MySQL更適合跨平台應用。 4)與MongoDB不同,MySQL更適用於結構化數據和事務處理。

MySQL通過表結構和SQL查詢高效管理結構化數據,並通過外鍵實現表間關係。 1.創建表時定義數據格式和類型。 2.使用外鍵建立表間關係。 3.通過索引和查詢優化提高性能。 4.定期備份和監控數據庫確保數據安全和性能優化。

MySQL值得學習,因為它是強大的開源數據庫管理系統,適用於數據存儲、管理和分析。 1)MySQL是關係型數據庫,使用SQL操作數據,適合結構化數據管理。 2)SQL語言是與MySQL交互的關鍵,支持CRUD操作。 3)MySQL的工作原理包括客戶端/服務器架構、存儲引擎和查詢優化器。 4)基本用法包括創建數據庫和表,高級用法涉及使用JOIN連接表。 5)常見錯誤包括語法錯誤和權限問題,調試技巧包括檢查語法和使用EXPLAIN命令。 6)性能優化涉及使用索引、優化SQL語句和定期維護數據庫。
