Skip to content
This repository was archived by the owner on Jan 6, 2023. It is now read-only.

Commit 990fbe4

Browse files
Kiuk Chungfacebook-github-bot
authored andcommitted
minor changes to README and github issue templates
Summary: see title Reviewed By: isunjin Differential Revision: D18838893 fbshipit-source-id: 17b888d8a6e1485ac36efed53c6bc22f6fb2c9d7
1 parent 93cdb7d commit 990fbe4

File tree

6 files changed

+55
-35
lines changed

6 files changed

+55
-35
lines changed

.github/ISSUE_TEMPLATE/bug-report.md

Lines changed: 20 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,21 @@ about: Submit a bug report to help us improve PyTorch Elastic
88

99
<!-- A clear and concise description of what the bug is. -->
1010

11+
Component (check all that applies):
12+
* [ ] `state api`
13+
* [ ] `train_step api`
14+
* [ ] `train_loop`
15+
* [ ] `rendezvous`
16+
* [ ] `checkpoint`
17+
* [ ] `rollback`
18+
* [ ] `metrics`
19+
* [ ] `petctl`
20+
* [ ] `examples`
21+
* [ ] `docker`
22+
* [ ] other
23+
24+
25+
1126
## To Reproduce
1227

1328
Steps to reproduce the behavior:
@@ -24,30 +39,17 @@ Steps to reproduce the behavior:
2439

2540
## Environment
2641

27-
- What commands did you use to install torchelastic (conda/pip/build from source)?
28-
- If you are building from source, which commit is it?
29-
- What does `classy_vision.__version__` print? (If applicable)
30-
31-
Please copy and paste the output from the Pytorch
32-
[environment collection script](https://raw.githubusercontent.com/pytorch/pytorch/master/torch/utils/collect_env.py)
33-
(or fill out the checklist below manually).
34-
35-
You can get the script and run it with:
36-
```
37-
wget https://raw.githubusercontent.com/pytorch/pytorch/master/torch/utils/collect_env.py
38-
# For security purposes, please check the contents of collect_env.py before running it.
39-
python collect_env.py
40-
```
41-
42-
- PyTorch Version (e.g., 1.0):
42+
- torchelastic version (e.g. 0.1.0rc1):
4343
- OS (e.g., Linux):
44-
- How you installed PyTorch (`conda`, `pip`, source):
44+
- How you installed torchelastic (`conda`, `pip`, source, `docker`):
45+
- Docker image and tag (if using docker):
4546
- Build command you used (if compiling from source):
47+
- Git commit (if installed from source):
4648
- Python version:
4749
- CUDA/cuDNN version:
4850
- GPU models and configuration:
49-
- Any other relevant information:
5051
- Execution environment (on-prem, aws, etc):
52+
- Any other relevant information:
5153

5254
## Additional context
5355

.github/ISSUE_TEMPLATE/feature-request.md

Lines changed: 11 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,24 +1,25 @@
11
---
22
name: "\U0001F680Feature Request"
3-
about: Submit a proposal/request for a new PyTorch Elastic feature
3+
about: Submit a proposal/request for a new feature or enhancement
44

55
---
66

7-
## 🚀 Feature
8-
<!-- A clear and concise description of the feature proposal -->
7+
## Description
8+
<!-- concise description of the feature/enhancement -->
99

10-
## Motivation
10+
## Motivation/Background
11+
<!-- why is this feature/enhancement important? provide background context -->
1112

12-
<!-- Please outline the motivation for the proposal. Is your feature request related to a problem? e.g., I'm always frustrated when [...]. If this is related to another GitHub issue, please link here too -->
1313

14-
## Pitch
14+
## Detailed Proposal
15+
<!-- provide a detailed proposal -->
1516

16-
<!-- A clear and concise description of what you want to happen. -->
1717

1818
## Alternatives
19+
<!-- discuss the alternativese considered and their pros/cons -->
1920

20-
<!-- A clear and concise description of any alternative solutions or features you've considered, if any. -->
2121

22-
## Additional context
22+
## Additional context/links
23+
<!-- link to code, documentation, etc -->
24+
2325

24-
<!-- Add any other context or screenshots about the feature request here. -->

.github/ISSUE_TEMPLATE/questions-help-support.md

Lines changed: 17 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,22 @@ about: Do you need support? We have resources.
66

77
## ❓ Questions and Help
88

9+
910
### Please note that this issue tracker is not a help form and this issue will be closed.
1011

11-
We have a set of [listed resources available on the website]. TODO: Fill this up.
12+
Before submitting, please ensure you have gone through our documentation. Here
13+
are some links that may be helpful:
14+
15+
* [What is torchelastic?](../../README.md)
16+
* [Quickstart on AWS](../../aws/README.md)
17+
* [Usage](../../USAGE.md)
18+
* [Examples](../../examples/README.md)
19+
* API documentation
20+
* [Overview](../../USAGE.md)
21+
* [Rendezvous documentation](../../torchelastic/rendezvous/README.md)
22+
* [Checkpointing documentation](../../torchelastic/checkpoint/README.md)
23+
* [Configuring](../../USAGE.md#configuring)
24+
25+
26+
### Question
27+
<!-- your question here -->

.github/PULL_REQUEST_TEMPLATE/pull_request_template.md

Whitespace-only changes.

README.md

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -105,12 +105,6 @@ in the rendezvous module.
105105
Please refer to the [usage documentation](USAGE.md) for details on how to write
106106
and configure a torchelastic job.
107107

108-
## Join the torchelastic community
109-
* Website:
110-
* Facebook page:
111-
* Mailing list
112-
* irc:
113-
114108
See the [CONTRIBUTING](CONTRIBUTING.md) file for how to help out.
115109

116110
## License

aws/README.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -140,6 +140,8 @@ Log into the AWS CloudWatch Logs console. You should see a log group called
140140
name `$job_name/$instance_id` (e.g. `my_job/i0b938EXAMPLE`).
141141

142142
#### Troubleshooting
143+
144+
##### SSH
143145
To SSH onto the worker nodes to debug/inspect the worker process use AWS
144146
Session Manager instead of the ec2 key pair. [Install](https://docs.aws.amazon.com/systems-manager/latest/userguide/session-manager-working-with-install-plugin.html)
145147
the Session Manager plugin and run
@@ -154,6 +156,7 @@ awscli ssm start-session --target <instance_id>
154156
awscli ssm start-session --target i-00b00EXAMPLE
155157
```
156158

159+
##### Process Status and Logs
157160
Once SSH'ed, the workers run in a docker container managed by `systemd`.
158161
You can take a look at their console outputs by running
159162

@@ -178,6 +181,10 @@ sudo systemctl start torchelastic_worker
178181
> **EXCERCISE:** Try stopping or adding worker(s) to see elasticity in action!
179182
To add workers, simply increase the `desired` size of the worker autoscaling group.
180183

184+
##### Notable Directories
185+
1. torchelastic_worker systemd: `/etc/systemd/service/torchelastic_worker.service`
186+
2. torchelastic run scripts: `/var/torchelastic`
187+
181188

182189
> **Note**: by design, `petctl` tries to use the least number of AWS services. This
183190
was done intentionally to allow non-AWS users to easily transfer the functionality

0 commit comments

Comments
 (0)