Skip to content

Commit

Permalink
Merge pull request #520 from GoogleCloudPlatform/release-v1.4.0
Browse files Browse the repository at this point in the history
Release v1.4.0
  • Loading branch information
nick-stroud authored Sep 7, 2022
2 parents 8525bf2 + 87f4b06 commit 6940ca6
Show file tree
Hide file tree
Showing 131 changed files with 1,341 additions and 978 deletions.
2 changes: 1 addition & 1 deletion cmd/root.go
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ HPC deployments on the Google Cloud Platform.`,
log.Fatalf("cmd.Help function failed: %s", err)
}
},
Version: "v1.3.0",
Version: "v1.4.0",
}
)

Expand Down
20 changes: 10 additions & 10 deletions community/examples/cloud-batch.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -27,19 +27,19 @@ vars:
deployment_groups:
- group: primary
modules:
- source: modules/network/pre-existing-vpc
- id: network1
source: modules/network/pre-existing-vpc
kind: terraform
id: network1

- source: modules/file-system/filestore
- id: appfs
source: modules/file-system/filestore
kind: terraform
id: appfs
use: [network1]
settings: {local_mount: /sw}

- source: modules/scripts/startup-script
- id: hello-startup-script
source: modules/scripts/startup-script
kind: terraform
id: hello-startup-script
settings:
runners:
- type: shell
Expand All @@ -53,9 +53,9 @@ deployment_groups:
#!/bin/sh
echo "Hello World" > /sw/hello.txt
- source: community/modules/scheduler/cloud-batch-job
- id: batch-job
source: community/modules/scheduler/cloud-batch-job
kind: terraform
id: batch-job
use: [network1, appfs, hello-startup-script]
settings:
runnable: "cat /sw/hello.txt"
Expand All @@ -64,8 +64,8 @@ deployment_groups:
family: centos-7
project: centos-cloud

- source: community/modules/scheduler/cloud-batch-login-node
- id: batch-login
source: community/modules/scheduler/cloud-batch-login-node
kind: terraform
id: batch-login
use: [batch-job]
outputs: [instructions]
24 changes: 12 additions & 12 deletions community/examples/hpc-cluster-small-sharedvpc.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -41,27 +41,27 @@ vars:
deployment_groups:
- group: primary
modules:
- source: modules/network/pre-existing-vpc
- id: network1
source: modules/network/pre-existing-vpc
kind: terraform
id: network1
settings:
project_id: $(vars.host_project_id)
network_name: your-shared-network
subnetwork_name: your-shared-subnetwork

- source: modules/file-system/filestore
- id: homefs
source: modules/file-system/filestore
kind: terraform
id: homefs
use: [network1]
settings:
local_mount: /home
connect_mode: PRIVATE_SERVICE_ACCESS
network_name: $(network1.network_id)

# This debug_partition will work out of the box without requesting additional GCP quota.
- source: community/modules/compute/SchedMD-slurm-on-gcp-partition
- id: debug_partition
source: community/modules/compute/SchedMD-slurm-on-gcp-partition
kind: terraform
id: debug_partition
use:
- network1
- homefs
Expand All @@ -73,19 +73,19 @@ deployment_groups:
machine_type: n2-standard-2

# This compute_partition is far more performant than debug_partition but may require requesting GCP quotas first.
- source: community/modules/compute/SchedMD-slurm-on-gcp-partition
- id: compute_partition
source: community/modules/compute/SchedMD-slurm-on-gcp-partition
kind: terraform
id: compute_partition
use:
- network1
- homefs
settings:
partition_name: compute
max_node_count: 20

- source: community/modules/scheduler/SchedMD-slurm-on-gcp-controller
- id: slurm_controller
source: community/modules/scheduler/SchedMD-slurm-on-gcp-controller
kind: terraform
id: slurm_controller
use:
- network1
- homefs
Expand All @@ -95,9 +95,9 @@ deployment_groups:
login_node_count: 1
shared_vpc_host_project: $(vars.host_project_id)

- source: community/modules/scheduler/SchedMD-slurm-on-gcp-login-node
- id: slurm_login
source: community/modules/scheduler/SchedMD-slurm-on-gcp-login-node
kind: terraform
id: slurm_login
use:
- network1
- homefs
Expand Down
40 changes: 20 additions & 20 deletions community/examples/htcondor-pool.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -27,32 +27,32 @@ vars:
deployment_groups:
- group: htcondor
modules:
- source: modules/network/vpc
- id: network1
source: modules/network/vpc
kind: terraform
id: network1
settings:
network_name: htcondor-pool
subnetwork_name: htcondor-pool-usc1
outputs:
- network_name

- source: community/modules/scripts/htcondor-install
- id: htcondor_install
source: community/modules/scripts/htcondor-install
kind: terraform
id: htcondor_install

- source: community/modules/project/service-enablement
- id: htcondor_services
source: community/modules/project/service-enablement
kind: terraform
id: htcondor_services
use:
- htcondor_install

- source: community/modules/scheduler/htcondor-configure
- id: htcondor_configure
source: community/modules/scheduler/htcondor-configure
kind: terraform
id: htcondor_configure

- source: modules/scripts/startup-script
- id: htcondor_configure_central_manager
source: modules/scripts/startup-script
kind: terraform
id: htcondor_configure_central_manager
settings:
runners:
- type: shell
Expand All @@ -61,9 +61,9 @@ deployment_groups:
- $(htcondor_install.install_htcondor_runner)
- $(htcondor_configure.central_manager_runner)

- source: modules/compute/vm-instance
- id: htcondor_cm
source: modules/compute/vm-instance
kind: terraform
id: htcondor_cm
use:
- network1
- htcondor_configure_central_manager
Expand All @@ -78,9 +78,9 @@ deployment_groups:
outputs:
- internal_ip

- source: modules/scripts/startup-script
- id: htcondor_configure_execute_point
source: modules/scripts/startup-script
kind: terraform
id: htcondor_configure_execute_point
settings:
runners:
- type: shell
Expand All @@ -89,9 +89,9 @@ deployment_groups:
- $(htcondor_install.install_htcondor_runner)
- $(htcondor_configure.execute_point_runner)

- source: community/modules/compute/htcondor-execute-point
- id: htcondor_execute_point
source: community/modules/compute/htcondor-execute-point
kind: terraform
id: htcondor_execute_point
use:
- network1
- htcondor_configure_execute_point
Expand All @@ -104,9 +104,9 @@ deployment_groups:
scopes:
- cloud-platform

- source: modules/scripts/startup-script
- id: htcondor_configure_access_point
source: modules/scripts/startup-script
kind: terraform
id: htcondor_configure_access_point
settings:
runners:
- type: shell
Expand All @@ -128,9 +128,9 @@ deployment_groups:
request_cpus = 1
request_memory = 100MB
queue
- source: modules/compute/vm-instance
- id: htcondor_access
source: modules/compute/vm-instance
kind: terraform
id: htcondor_access
use:
- network1
- htcondor_configure_access_point
Expand Down
12 changes: 6 additions & 6 deletions community/examples/intel/daos-cluster.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -28,16 +28,16 @@ vars:
deployment_groups:
- group: primary
modules:
- source: modules/network/pre-existing-vpc
- id: network1
source: modules/network/pre-existing-vpc
kind: terraform
id: network1

# This module creates a DAOS server. Server images MUST be created before running this.
# https://github.com/daos-stack/google-cloud-daos/tree/main/images
# more info: https://github.com/daos-stack/google-cloud-daos/tree/main/terraform/modules/daos_server
- source: github.com/daos-stack/google-cloud-daos.git//terraform/modules/daos_server?ref=v0.2.1
- id: daos-server
source: github.com/daos-stack/google-cloud-daos.git//terraform/modules/daos_server?ref=v0.2.1
kind: terraform
id: daos-server
use: [network1]
settings:
number_of_instances: 2
Expand All @@ -46,9 +46,9 @@ deployment_groups:
# This module creates a MIG with DAOS clients. Client images MUST be created before running this.
# https://github.com/daos-stack/google-cloud-daos/tree/main/images
# more info: https://github.com/daos-stack/google-cloud-daos/tree/main/terraform/modules/daos_client
- source: github.com/daos-stack/google-cloud-daos.git//terraform/modules/daos_client?ref=v0.2.1
- id: daos-client
source: github.com/daos-stack/google-cloud-daos.git//terraform/modules/daos_client?ref=v0.2.1
kind: terraform
id: daos-client
use: [network1, daos-server]
settings:
number_of_instances: 2
Expand Down
32 changes: 16 additions & 16 deletions community/examples/intel/daos-slurm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -28,23 +28,23 @@ vars:
deployment_groups:
- group: primary
modules:
- source: modules/network/pre-existing-vpc
- id: network1
source: modules/network/pre-existing-vpc
kind: terraform
id: network1

- source: modules/file-system/filestore
- id: homefs
source: modules/file-system/filestore
kind: terraform
id: homefs
use: [network1]
settings:
local_mount: "/home"

# This module creates a DAOS server. Server images MUST be created before running this.
# https://github.com/daos-stack/google-cloud-daos/tree/main/images
# more info: https://github.com/daos-stack/google-cloud-daos/tree/main/terraform/modules/daos_server
- source: github.com/daos-stack/google-cloud-daos.git//terraform/modules/daos_server?ref=v0.2.1
- id: daos
source: github.com/daos-stack/google-cloud-daos.git//terraform/modules/daos_server?ref=v0.2.1
kind: terraform
id: daos
use: [network1]
settings:
labels: {ghpc_role: file-system}
Expand All @@ -68,9 +68,9 @@ deployment_groups:
reclaim: "lazy"
containers: []

- source: modules/scripts/startup-script
- id: daos-client-script
source: modules/scripts/startup-script
kind: terraform
id: daos-client-script
settings:
runners:
- type: shell
Expand All @@ -87,9 +87,9 @@ deployment_groups:
destination: /var/daos/daos_client_config.sh

## This debug_partition will work out of the box without requesting additional GCP quota.
- source: community/modules/compute/SchedMD-slurm-on-gcp-partition
- id: debug_partition
source: community/modules/compute/SchedMD-slurm-on-gcp-partition
kind: terraform
id: debug_partition
use:
- network1
- homefs
Expand All @@ -100,19 +100,19 @@ deployment_groups:
machine_type: n2-standard-2

# This compute_partition is far more performant than debug_partition but may require requesting GCP quotas first.
- source: community/modules/compute/SchedMD-slurm-on-gcp-partition
- id: compute_partition
source: community/modules/compute/SchedMD-slurm-on-gcp-partition
kind: terraform
id: compute_partition
use:
- network1
- homefs
settings:
partition_name: compute
max_node_count: 20

- source: community/modules/scheduler/SchedMD-slurm-on-gcp-controller
- id: slurm_controller
source: community/modules/scheduler/SchedMD-slurm-on-gcp-controller
kind: terraform
id: slurm_controller
use:
- network1
- homefs
Expand All @@ -127,9 +127,9 @@ deployment_groups:
- "https://www.googleapis.com/auth/devstorage.read_only"
- "https://www.googleapis.com/auth/cloud-platform"

- source: community/modules/scheduler/SchedMD-slurm-on-gcp-login-node
- id: slurm_login
source: community/modules/scheduler/SchedMD-slurm-on-gcp-login-node
kind: terraform
id: slurm_login
use:
- network1
- homefs
Expand Down
Loading

0 comments on commit 6940ca6

Please sign in to comment.