Fix ansible-lint errors
This commit is contained in:
		
							parent
							
								
									503bd8b396
								
							
						
					
					
						commit
						b098704e23
					
				
					 10 changed files with 21 additions and 26 deletions
				
			
		|  | @ -3,12 +3,12 @@ | |||
|   hosts: all | ||||
|   become: true | ||||
|   tasks: | ||||
|   - name: Get new password | ||||
|     ansible.builtin.pause: | ||||
|       prompt: Enter the new password | ||||
|       echo: false | ||||
|     register: new_user_password | ||||
|   - name: Change password | ||||
|     ansible.builtin.user: | ||||
|       name: "{{ interactive_user }}" | ||||
|       password: "{{ new_user_password.user_input | password_hash('sha512') }}" | ||||
|     - name: Get new password | ||||
|       ansible.builtin.pause: | ||||
|         prompt: Enter the new password | ||||
|         echo: false | ||||
|       register: new_user_password | ||||
|     - name: Change password | ||||
|       ansible.builtin.user: | ||||
|         name: "{{ interactive_user }}" | ||||
|         password: "{{ new_user_password.user_input | password_hash('sha512') }}" | ||||
|  |  | |||
|  | @ -1,3 +0,0 @@ | |||
| --- | ||||
| - name: Restart service {{ role_name }} | ||||
|   ansible.builtin.include_tasks: tasks/restart-service.yml | ||||
|  | @ -1,12 +0,0 @@ | |||
| --- | ||||
| - name: Set common facts | ||||
|   ansible.builtin.import_tasks: tasks/set-default-facts.yml | ||||
| 
 | ||||
| - name: Deploy {{ role_name }} | ||||
|   vars: | ||||
|     svc: "{{ NAME_svc }}" | ||||
|     env: "{{ NAME_env }}" | ||||
|     compose: "{{ NAME_compose }}" | ||||
|   block: | ||||
|     - name: Import tasks to deploy common service | ||||
|       ansible.builtin.import_tasks: tasks/deploy-common-service.yml | ||||
|  | @ -1,16 +0,0 @@ | |||
| --- | ||||
| NAME_svc: | ||||
|   domain: "{{ all_services | service_get_domain(role_name) }}" | ||||
|   port: 80 | ||||
| 
 | ||||
| NAME_env: | ||||
|   EXAMPLE: value | ||||
| 
 | ||||
| NAME_compose: | ||||
|   watchtower: update | ||||
|   image:  | ||||
|   volumes: | ||||
|     - data:/data | ||||
|   file: | ||||
|     volumes: | ||||
|       data: | ||||
|  | @ -30,4 +30,4 @@ deploy_compose: | |||
|   file: | ||||
|     services: | ||||
|       app: | ||||
|         command: ["-verbose", "-hooks=/config/hooks.yml", "-hotreload"] | ||||
|         command: ["-verbose", "-hooks=/config/hooks.yml", "-hotreload"] | ||||
|  |  | |||
|  | @ -7,13 +7,6 @@ runner: | |||
|   file: /data/.runner | ||||
|   # Execute how many tasks concurrently at the same time. | ||||
|   capacity: 1 | ||||
|   # Extra environment variables to run jobs. | ||||
|   #envs: | ||||
|   #  A_TEST_ENV_NAME_1: a_test_env_value_1 | ||||
|   #  A_TEST_ENV_NAME_2: a_test_env_value_2 | ||||
|   # Extra environment variables to run jobs from a file. | ||||
|   # It will be ignored if it's empty or the file doesn't exist. | ||||
|   #env_file: .env | ||||
|   # The timeout for a job to be finished. | ||||
|   # Please note that the Forgejo instance also has a timeout (3h by default) for the job. | ||||
|   # So the job could be stopped by the Forgejo instance if it's timeout is shorter than this. | ||||
|  |  | |||
|  | @ -1,4 +1,4 @@ | |||
| --- | ||||
| - name: Deploy {{ role_name }} | ||||
|   ansible.builtin.import_role: | ||||
|     name: prometheus.prometheus.node_exporter | ||||
|     name: prometheus.prometheus.node_exporter | ||||
|  |  | |||
|  | @ -3,7 +3,8 @@ | |||
|   hosts: serguzim_net | ||||
|   become: true | ||||
|   tasks: | ||||
|   - name: Unlock backups | ||||
|     ansible.builtin.shell: | ||||
|       cmd: autorestic unlock --force && autorestic exec -va unlock | ||||
|       chdir: "{{ (services_path, 'backup') | path_join }}" | ||||
|     - name: Unlock backups | ||||
|       ansible.builtin.shell: | ||||
|         cmd: autorestic unlock --force && autorestic exec -va unlock | ||||
|         chdir: "{{ (services_path, 'backup') | path_join }}" | ||||
|       changed_when: true | ||||
|  |  | |||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue