Compare commits
309 Commits
Author | SHA1 | Date | |
---|---|---|---|
ebda4660de | |||
f44e377e29 | |||
0a6d4c998d | |||
bd9733d929 | |||
9ae3282dcf | |||
e62fedbd5e | |||
c53ae2afa0 | |||
45e7f6a30c | |||
8fbedf2886 | |||
c0063a6573 | |||
c9795a8213 | |||
8aec402ea2 | |||
adc2079b17 | |||
cee1b5b079 | |||
13d66f2ae7 | |||
901ac05e99 | |||
2bdaed1e6f | |||
b45934cd9e | |||
fceb6e0e13 | |||
f05a9c67ae | |||
f2d1e62204 | |||
8f704f220a | |||
d5093201ee | |||
401777adff | |||
f80cc70d76 | |||
5b8bf780df | |||
bce52596a5 | |||
c3c94b0a63 | |||
14c30fb81c | |||
1130ff6c6d | |||
993c93b34a | |||
dcf0bdb950 | |||
9e2f4313fb | |||
df41020cb8 | |||
1332df4857 | |||
3f23419bb7 | |||
184ff72912 | |||
55ee95df78 | |||
78fe47aaba | |||
1370596f27 | |||
606e56d78f | |||
51226c8e50 | |||
6c9d4aebac | |||
ed8b42fbde | |||
aa56f645e9 | |||
9a1bb788d2 | |||
7d32f79379 | |||
0410cf2fd2 | |||
c9f5e34c0d | |||
047f2abdb5 | |||
f13a6573a8 | |||
596bb6953c | |||
d482e623df | |||
bf910dea02 | |||
57da852af6 | |||
4c21344e8b | |||
302c0fed59 | |||
0e7afe466d | |||
2c757af250 | |||
27f28935d7 | |||
12753dd7d4 | |||
dd8dce1b49 | |||
19280c2bb0 | |||
5c4a1cc082 | |||
a63af05bec | |||
9be7b0e109 | |||
d5b1d9ce39 | |||
a6eb4fef00 | |||
d1c69e8fe5 | |||
81d9e4f560 | |||
61a3afbb3b | |||
91ab616cc5 | |||
43d62f1284 | |||
d61470d6ef | |||
091d23d537 | |||
2f24902436 | |||
36d47b9b88 | |||
2c20771682 | |||
a9cdec6d87 | |||
f1efccea6b | |||
226d8438de | |||
b6d6f85dc0 | |||
49ed10bbee | |||
3afdd3bf13 | |||
0170ed7d6f | |||
dad2e93572 | |||
d726426611 | |||
b31d99b936 | |||
8860d32f97 | |||
b9dff0d22e | |||
297aab1fd3 | |||
54bf7a3819 | |||
52876be723 | |||
ff20e4f49c | |||
ee47ba04bc | |||
688374d13f | |||
22c4212438 | |||
26dcb2f2e0 | |||
46afa266fa | |||
b2d26d9dce | |||
0b622361f3 | |||
ed48909f96 | |||
60d191bb2a | |||
4f482e484f | |||
58a1579832 | |||
6d5cf7a815 | |||
2e92554423 | |||
2456a642b7 | |||
b5108625ff | |||
9ab54f27b1 | |||
4b7037bbe6 | |||
f183603bc0 | |||
544a58325d | |||
9557bdf209 | |||
4890a8510d | |||
66f77b5eb1 | |||
661b6b07a3 | |||
ce8db1065d | |||
9a0e1115ed | |||
2b4c058af8 | |||
a206ad6811 | |||
a1dabcb758 | |||
940c1bb50c | |||
6c49d4fef5 | |||
621df33df9 | |||
7e86ba002f | |||
6f12cb36af | |||
4ef35aadb6 | |||
d4081c15e1 | |||
6e870b4d0b | |||
1e78ccd204 | |||
84e9b63524 | |||
4cc7f13b6e | |||
fc2d0d1852 | |||
ef98ad54fa | |||
4785e09945 | |||
d4f9bb6da3 | |||
ab41eaa5d9 | |||
02794f368b | |||
9c66cc4c32 | |||
68401caf1c | |||
4f466bb5a1 | |||
daaea77144 | |||
faa1cb52dd | |||
57ac14f295 | |||
d0f4820baa | |||
b0261e8cc6 | |||
30a7aa0e9c | |||
5f0b7467d1 | |||
24363b653f | |||
21a3048b96 | |||
5679ab435f | |||
c9b3edd0f1 | |||
31e6e42a30 | |||
1a868c6847 | |||
a9d6f1fefe | |||
dbfcd5b58c | |||
6a52a48b71 | |||
e31492a9b4 | |||
f2dbc4f717 | |||
1d040ad64d | |||
8703749292 | |||
d432a84c42 | |||
c00492cad0 | |||
c4d545a256 | |||
d0c429c8a1 | |||
a8b6883183 | |||
5590b32c93 | |||
97e7022121 | |||
2926d1a11d | |||
58debafa8a | |||
e021390568 | |||
389e68f6ba | |||
7eeb2ea7e6 | |||
954bddeb64 | |||
9255afd060 | |||
a0a32b930e | |||
fc4b78f74a | |||
9e063a6e6f | |||
54823b41e1 | |||
6c5434ee20 | |||
8e0c795791 | |||
1fd8d105e6 | |||
fe4df4f70a | |||
650e62f44f | |||
f8451b0c0a | |||
3af075e039 | |||
26d45b1e85 | |||
9b6c476bf9 | |||
c5a1557800 | |||
9d476fa9af | |||
9142aaf78e | |||
b258b6578c | |||
289540f576 | |||
26a3f5fcc0 | |||
18ea220102 | |||
03a1793208 | |||
3c01985e20 | |||
275d2673b4 | |||
746aa9b4a6 | |||
c13bec47b8 | |||
9953d11fd8 | |||
b2ee468871 | |||
dd41d09b43 | |||
abfa6c989f | |||
bb3d964b89 | |||
6bbc7a3a84 | |||
17b1a5022d | |||
e38d646e8e | |||
bb298e7a01 | |||
98840382b9 | |||
b8ff0a1d43 | |||
15b097bd3e | |||
e0b34b8c67 | |||
3aad44d68c | |||
6658d069fb | |||
52f097d4a2 | |||
f4e9ebcfb5 | |||
057728b774 | |||
5b618851c0 | |||
9f531ce0f7 | |||
dd732d7bc2 | |||
05bb40adb1 | |||
240326ef37 | |||
c49fa313ed | |||
ff6bd95fce | |||
0def662e53 | |||
49b7c267b9 | |||
02f30f3a68 | |||
e2cc9386fb | |||
43b8b36951 | |||
4d320f34d5 | |||
676ca798cd | |||
60428d5639 | |||
fd55ec220c | |||
73b07a7181 | |||
1f3d9463f9 | |||
8678ad92ce | |||
5b34df1471 | |||
3d063a4b35 | |||
f924e76adf | |||
604ff3b76d | |||
96545135cb | |||
474ca9e78f | |||
12401f2a3f | |||
528bc17eb8 | |||
4e31218ce6 | |||
d1b70ba2d4 | |||
ff0bc9757a | |||
a1e0c52825 | |||
45889c3811 | |||
71d8ac10ac | |||
744ed91e2a | |||
b1bcfcbc4e | |||
a7a5d13637 | |||
5d4ecb7f9e | |||
69dd87689c | |||
9fab9a3434 | |||
b69a97d2ca | |||
8176ef8499 | |||
eb5cc88b25 | |||
f2731d3fe6 | |||
6f4d025cb7 | |||
222ede2c4b | |||
5304c8f901 | |||
28a26288c9 | |||
53d759e263 | |||
6879412e2b | |||
a89d2e035c | |||
d5c92e9a7d | |||
c7df608696 | |||
eee07f703a | |||
6e7d5a6f59 | |||
752f592e58 | |||
f637de88f0 | |||
a5da38203f | |||
14c0d17704 | |||
a572d5f5de | |||
3673196c16 | |||
9874cef369 | |||
1c4b6741ac | |||
b18e4aaf92 | |||
45824ba4f3 | |||
5944086435 | |||
a2e33c166d | |||
9ab202f921 | |||
cca79b5dbd | |||
61eecdd343 | |||
ce6fe5cb83 | |||
8f119ede13 | |||
edb61f3d90 | |||
cb5ffa0354 | |||
ff1eed3cf0 | |||
53d593de25 | |||
8111da308b | |||
0125ab4849 | |||
ad0c04985b | |||
793393fe0f | |||
59995a8719 | |||
f19d650412 | |||
6cd156f7d5 | |||
95976f4cf4 | |||
0dec4d8df4 | |||
22186b267d | |||
020e0adeb4 | |||
009eb1fc6d | |||
989c250a82 | |||
3ee231bdd8 | |||
3338cdc9d7 |
67
.github/ISSUE_TEMPLATE/bug_report.md
vendored
67
.github/ISSUE_TEMPLATE/bug_report.md
vendored
@ -1,67 +0,0 @@
|
||||
---
|
||||
name: Bug report
|
||||
about: Create a report to help us improve
|
||||
title: ''
|
||||
labels: ''
|
||||
assignees: ''
|
||||
|
||||
---
|
||||
|
||||
<!--
|
||||
|
||||
Before raising an issue here, answer the following questions for yourself, please:
|
||||
|
||||
* Did you read through the troubleshooting section? (https://github.com/netbox-community/netbox-docker/wiki/Troubleshooting)
|
||||
* Have you had a look at the rest of the wiki? (https://github.com/netbox-community/netbox-docker/wiki)
|
||||
* Have you updated to the latest version and tried again? (i.e. `git pull` and `docker-compose pull`)
|
||||
* Have you reset the project and tried again? (i.e. `docker-compose down -v`)
|
||||
* Are you confident that your problem is related to the Docker image or Docker Compose file this project provides?
|
||||
(Otherwise ask on the NetBox mailing list, please: https://groups.google.com/d/forum/netbox-discuss)
|
||||
* Have you looked through the issues already resolved?
|
||||
|
||||
Please try this means to get help before opening an issue here:
|
||||
|
||||
* On the networktocode Slack in the #netbox-docker channel: http://slack.networktocode.com/
|
||||
* On the networktocode Slack in the #netbox channel: http://slack.networktocode.com/
|
||||
* On the NetBox mailing list: https://groups.google.com/d/forum/netbox-discuss
|
||||
|
||||
Please don't open an issue when you have a PR ready. Just submit the PR, that's good enough.
|
||||
|
||||
-->
|
||||
|
||||
## Current Behavior
|
||||
|
||||
<!-- describe what you did and how it misbehaved -->
|
||||
...
|
||||
|
||||
## Expected Behavior
|
||||
|
||||
<!-- describe what you expected instead -->
|
||||
...
|
||||
|
||||
## Debug Information
|
||||
|
||||
<!-- please fill in the following information that might helps us debug your problem more quickly -->
|
||||
The output of `docker-compose version`: `XXXXX`
|
||||
The output of `docker version`: `XXXXX`
|
||||
The output of `git rev-parse HEAD`: `XXXXX`
|
||||
The command you used to start the project: `XXXXX`
|
||||
|
||||
<!-- adjust the `latest` tag to the version you're using -->
|
||||
The output of `docker inspect netboxcommunity/netbox:latest --format "{{json .Config.Labels}}"`:
|
||||
|
||||
```json
|
||||
{
|
||||
"JSON JSON JSON":
|
||||
"--> Please paste formatted json. (Use e.g. `jq` or https://jsonformatter.curiousconcept.com/)"
|
||||
}
|
||||
```
|
||||
|
||||
The output of `docker-compose logs netbox`:
|
||||
<!--
|
||||
If your log is very long, create a Gist instead (and post the link to it): https://gist.github.com
|
||||
-->
|
||||
|
||||
```text
|
||||
LOG LOG LOG
|
||||
```
|
148
.github/ISSUE_TEMPLATE/bug_report.yml
vendored
Normal file
148
.github/ISSUE_TEMPLATE/bug_report.yml
vendored
Normal file
@ -0,0 +1,148 @@
|
||||
name: Bug report
|
||||
description: Create a report about a malfunction of the Docker setup
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: |
|
||||
Please only raise an issue if you're certain that you've found a bug.
|
||||
Else, see these other means to get help:
|
||||
|
||||
- See our troubleshooting section:
|
||||
https://github.com/netbox-community/netbox-docker/wiki/Troubleshooting
|
||||
- Have a look at the rest of the wiki:
|
||||
https://github.com/netbox-community/netbox-docker/wiki
|
||||
- Check the release notes:
|
||||
https://github.com/netbox-community/netbox-docker/releases
|
||||
- Look through the issues already resolved:
|
||||
https://github.com/netbox-community/netbox-docker/issues?q=is%3Aclosed
|
||||
|
||||
If you did not find what you're looking for,
|
||||
try the help of our community:
|
||||
|
||||
- Post to Github Discussions:
|
||||
https://github.com/netbox-community/netbox-docker/discussions
|
||||
- Join the `#netbox-docker` channel on our Slack:
|
||||
https://join.slack.com/t/netdev-community/shared_invite/zt-mtts8g0n-Sm6Wutn62q_M4OdsaIycrQ
|
||||
- Ask on the NetBox mailing list:
|
||||
https://groups.google.com/d/forum/netbox-discuss
|
||||
|
||||
Please don't open an issue to open a PR.
|
||||
Just submit the PR, that's good enough.
|
||||
- type: textarea
|
||||
id: current-behavior
|
||||
attributes:
|
||||
label: Current Behavior
|
||||
description: Please describe what you did and how you think it misbehaved
|
||||
placeholder: I tried to … by doing …, but it …
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
id: expected-behavior
|
||||
attributes:
|
||||
label: Expected Behavior
|
||||
description: Please describe what you expected instead
|
||||
placeholder: I expected that … when I do …
|
||||
validations:
|
||||
required: true
|
||||
- type: input
|
||||
id: docker-compose-version
|
||||
attributes:
|
||||
label: Docker Compose Version
|
||||
description: Please paste the output of `docker-compose version`
|
||||
placeholder: Docker Compose version vX.Y.Z
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
id: docker-version
|
||||
attributes:
|
||||
label: Docker Version
|
||||
description: Please paste the output of `docker version`
|
||||
render: text
|
||||
placeholder: |
|
||||
Client:
|
||||
Cloud integration: 1.0.17
|
||||
Version: 20.10.8
|
||||
API version: 1.41
|
||||
Go version: go1.16.6
|
||||
Git commit: 3967b7d
|
||||
Built: Fri Jul 30 19:55:20 2021
|
||||
OS/Arch: darwin/amd64
|
||||
Context: default
|
||||
Experimental: true
|
||||
|
||||
Server: Docker Engine - Community
|
||||
Engine:
|
||||
Version: 20.10.8
|
||||
API version: 1.41 (minimum version 1.12)
|
||||
Go version: go1.16.6
|
||||
Git commit: 75249d8
|
||||
Built: Fri Jul 30 19:52:10 2021
|
||||
OS/Arch: linux/amd64
|
||||
Experimental: false
|
||||
containerd:
|
||||
Version: 1.4.9
|
||||
GitCommit: e25210fe30a0a703442421b0f60afac609f950a3
|
||||
runc:
|
||||
Version: 1.0.1
|
||||
GitCommit: v1.0.1-0-g4144b63
|
||||
docker-init:
|
||||
Version: 0.19.0
|
||||
GitCommit: de40ad0
|
||||
validations:
|
||||
required: true
|
||||
- type: input
|
||||
id: git-rev
|
||||
attributes:
|
||||
label: The git Revision
|
||||
description: Please paste the output of `git rev-parse HEAD`
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
id: git-status
|
||||
attributes:
|
||||
label: The git Status
|
||||
description: Please paste the output of `git status`
|
||||
render: text
|
||||
placeholder: |
|
||||
On branch main
|
||||
nothing to commit, working tree clean
|
||||
validations:
|
||||
required: true
|
||||
- type: input
|
||||
id: run-command
|
||||
attributes:
|
||||
label: Startup Command
|
||||
description: Please specify the command you used to start the project
|
||||
placeholder: docker compose up
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
id: netbox-logs
|
||||
attributes:
|
||||
label: NetBox Logs
|
||||
description: Please paste the output of `docker-compose logs netbox` (or `docker compose logs netbox`)
|
||||
render: text
|
||||
placeholder: |
|
||||
netbox_1 | ⚙️ Applying database migrations
|
||||
netbox_1 | 🧬 loaded config '/etc/netbox/config/configuration.py'
|
||||
netbox_1 | 🧬 loaded config '/etc/netbox/config/a.py'
|
||||
netbox_1 | 🧬 loaded config '/etc/netbox/config/extra.py'
|
||||
netbox_1 | 🧬 loaded config '/etc/netbox/config/logging.py'
|
||||
netbox_1 | 🧬 loaded config '/etc/netbox/config/plugins.py'
|
||||
...
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
id: docker-compose-override-yml
|
||||
attributes:
|
||||
label: Content of docker-compose.override.yml
|
||||
description: Please paste the output of `cat docker-compose.override.yml`
|
||||
render: yaml
|
||||
placeholder: |
|
||||
version: '3.4'
|
||||
services:
|
||||
netbox:
|
||||
ports:
|
||||
- '8080:8080'
|
||||
validations:
|
||||
required: true
|
2
.github/ISSUE_TEMPLATE/config.yml
vendored
2
.github/ISSUE_TEMPLATE/config.yml
vendored
@ -5,7 +5,7 @@ contact_links:
|
||||
about: The Github Discussions are the right place to ask questions about how to use or do certain things with NetBox Docker.
|
||||
|
||||
- name: Chat
|
||||
url: http://slack.networktocode.com/
|
||||
url: https://join.slack.com/t/netdev-community/shared_invite/zt-mtts8g0n-Sm6Wutn62q_M4OdsaIycrQ
|
||||
about: 'Usually the quickest way to seek help with small issues is to join our #netbox-docker Slack channel.'
|
||||
|
||||
- name: Community Wiki
|
||||
|
54
.github/ISSUE_TEMPLATE/feature_request.md
vendored
54
.github/ISSUE_TEMPLATE/feature_request.md
vendored
@ -1,54 +0,0 @@
|
||||
---
|
||||
name: Feature or Change Request
|
||||
about: Request a new feature or a change of the current behavior
|
||||
title: ''
|
||||
labels: ''
|
||||
assignees: ''
|
||||
|
||||
---
|
||||
|
||||
<!--
|
||||
|
||||
Before raising an issue here, answer the following questions for yourself, please:
|
||||
|
||||
* Did you read through the troubleshooting section? (https://github.com/netbox-community/netbox-docker/wiki/Troubleshooting)
|
||||
* Have you had a look at the rest of the wiki? (https://github.com/netbox-community/netbox-docker/wiki)
|
||||
* Have you read the release notes recently (https://github.com/netbox-community/netbox-docker/releases)
|
||||
* Are you confident that your feature/change request is related to the Docker image or Docker Compose file this project provides?
|
||||
(Otherwise ask on the NetBox mailing list, please: https://groups.google.com/d/forum/netbox-discuss)
|
||||
* Have you looked through the issues already resolved?
|
||||
|
||||
Please try this means to get help before opening an issue here:
|
||||
|
||||
* On the networktocode Slack in the #netbox-docker channel: http://slack.networktocode.com/
|
||||
* On the networktocode Slack in the #netbox channel: http://slack.networktocode.com/
|
||||
* On the NetBox mailing list: https://groups.google.com/d/forum/netbox-discuss
|
||||
|
||||
Please don't open an issue when you have a PR ready. Just submit the PR, that's good enough.
|
||||
|
||||
-->
|
||||
|
||||
## Desired Behavior
|
||||
|
||||
<!-- please describe the behavior you desire -->
|
||||
...
|
||||
|
||||
## Contrast to Current Behavior
|
||||
|
||||
<!-- please describe how the desired behavior is different from the current behavior -->
|
||||
...
|
||||
|
||||
## Changes Required
|
||||
|
||||
<!-- if you can, please elaborate what changes would exactly be required -->
|
||||
...
|
||||
|
||||
## Discussion: Benefits and Drawbacks
|
||||
|
||||
<!--
|
||||
Please make your case here:
|
||||
- Why do you think this project and the community will benefit from your suggestion?
|
||||
- What are the drawbacks of this change? Is it backwards-compatible?
|
||||
- Anything else that you think is relevant to the discussion of this feature/change request.
|
||||
-->
|
||||
...
|
68
.github/ISSUE_TEMPLATE/feature_request.yml
vendored
Normal file
68
.github/ISSUE_TEMPLATE/feature_request.yml
vendored
Normal file
@ -0,0 +1,68 @@
|
||||
name: Feature or Change Request
|
||||
description: Request a new feature or a change of the current behavior
|
||||
body:
|
||||
- type: markdown
|
||||
attributes:
|
||||
value: |
|
||||
This issue type is to propose new features for the Docker setup.
|
||||
To just spin an idea, see the Github Discussions section, please.
|
||||
|
||||
Before asking for help, see these links first:
|
||||
|
||||
- See our troubleshooting section:
|
||||
https://github.com/netbox-community/netbox-docker/wiki/Troubleshooting
|
||||
- Have a look at the rest of the wiki:
|
||||
https://github.com/netbox-community/netbox-docker/wiki
|
||||
- Check the release notes:
|
||||
https://github.com/netbox-community/netbox-docker/releases
|
||||
- Look through the issues already resolved:
|
||||
https://github.com/netbox-community/netbox-docker/issues?q=is%3Aclosed
|
||||
|
||||
If you did not find what you're looking for,
|
||||
try the help of our community:
|
||||
|
||||
- Post to Github Discussions:
|
||||
https://github.com/netbox-community/netbox-docker/discussions
|
||||
- Join the `#netbox-docker` channel on our Slack:
|
||||
https://join.slack.com/t/netdev-community/shared_invite/zt-mtts8g0n-Sm6Wutn62q_M4OdsaIycrQ
|
||||
- Ask on the NetBox mailing list:
|
||||
https://groups.google.com/d/forum/netbox-discuss
|
||||
|
||||
Please don't open an issue to open a PR.
|
||||
Just submit the PR, that's good enough.
|
||||
- type: textarea
|
||||
id: desired-behavior
|
||||
attributes:
|
||||
label: Desired Behavior
|
||||
description: Please describe the desired behavior
|
||||
placeholder: To me, it would be useful, if … because …
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
id: contrast-to-current
|
||||
attributes:
|
||||
label: Contrast to Current Behavior
|
||||
description: Please describe how the desired behavior is different from the current behavior
|
||||
placeholder: The current behavior is …, but this lacks …
|
||||
validations:
|
||||
required: true
|
||||
- type: textarea
|
||||
id: required-changes
|
||||
attributes:
|
||||
label: Required Changes
|
||||
description: If you can, please elaborate what changes will be required to implement the desired behavior
|
||||
placeholder: I suggest to change the file …
|
||||
validations:
|
||||
required: false
|
||||
- type: textarea
|
||||
id: discussion
|
||||
attributes:
|
||||
label: 'Discussion: Benefits and Drawbacks'
|
||||
description: |
|
||||
Please make your case here:
|
||||
- Why do you think this project and the community will benefit from your suggestion?
|
||||
- What are the drawbacks of this change? Is it backwards-compatible?
|
||||
- Anything else that you think is relevant to the discussion of this feature/change request.
|
||||
placeholder: I suggest to change the file …
|
||||
validations:
|
||||
required: false
|
96
.github/workflows/push.yml
vendored
96
.github/workflows/push.yml
vendored
@ -1,3 +1,4 @@
|
||||
---
|
||||
name: push
|
||||
|
||||
on:
|
||||
@ -13,51 +14,66 @@ jobs:
|
||||
runs-on: ubuntu-latest
|
||||
name: Checks syntax of our code
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- uses: actions/setup-python@v2
|
||||
- name: Lint Code Base
|
||||
uses: github/super-linter@v3
|
||||
env:
|
||||
DEFAULT_BRANCH: develop
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
SUPPRESS_POSSUM: true
|
||||
LINTER_RULES_PATH: /
|
||||
VALIDATE_ALL_CODEBASE: false
|
||||
VALIDATE_DOCKERFILE: false
|
||||
FILTER_REGEX_EXCLUDE: (.*/)?(LICENSE|configuration/.*)
|
||||
|
||||
EDITORCONFIG_FILE_NAME: .ecrc
|
||||
DOCKERFILE_HADOLINT_FILE_NAME: .hadolint.yaml
|
||||
MARKDOWN_CONFIG_FILE: .markdown-lint.yml
|
||||
PYTHON_BLACK_CONFIG_FILE: pyproject.toml
|
||||
PYTHON_FLAKE8_CONFIG_FILE: .flake8
|
||||
PYTHON_ISORT_CONFIG_FILE: pyproject.toml
|
||||
- uses: actions/checkout@v3
|
||||
with:
|
||||
# Full git history is needed to get a proper
|
||||
# list of changed files within `super-linter`
|
||||
fetch-depth: 0
|
||||
- uses: actions/setup-python@v4
|
||||
with:
|
||||
python-version: '3.9'
|
||||
- name: Lint Code Base
|
||||
uses: github/super-linter@v4
|
||||
env:
|
||||
DEFAULT_BRANCH: develop
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
SUPPRESS_POSSUM: true
|
||||
LINTER_RULES_PATH: /
|
||||
VALIDATE_ALL_CODEBASE: false
|
||||
VALIDATE_DOCKERFILE: false
|
||||
FILTER_REGEX_EXCLUDE: (.*/)?(LICENSE|configuration/.*)
|
||||
EDITORCONFIG_FILE_NAME: .ecrc
|
||||
DOCKERFILE_HADOLINT_FILE_NAME: .hadolint.yaml
|
||||
MARKDOWN_CONFIG_FILE: .markdown-lint.yml
|
||||
PYTHON_BLACK_CONFIG_FILE: pyproject.toml
|
||||
PYTHON_FLAKE8_CONFIG_FILE: .flake8
|
||||
PYTHON_ISORT_CONFIG_FILE: pyproject.toml
|
||||
YAML_CONFIG_FILE: .yamllint.yaml
|
||||
build:
|
||||
continue-on-error: ${{ matrix.docker_from == 'alpine:edge' }}
|
||||
continue-on-error: ${{ matrix.build_cmd != './build-latest.sh' }}
|
||||
strategy:
|
||||
matrix:
|
||||
build_cmd:
|
||||
- ./build-latest.sh
|
||||
- PRERELEASE=true ./build-latest.sh
|
||||
- ./build-next.sh
|
||||
- ./build.sh develop
|
||||
docker_from:
|
||||
- '' # use the default of the build script
|
||||
- alpine:edge
|
||||
- ./build-latest.sh
|
||||
- PRERELEASE=true ./build-latest.sh
|
||||
- ./build.sh feature
|
||||
- ./build.sh develop
|
||||
platform:
|
||||
- linux/amd64
|
||||
- linux/arm64
|
||||
fail-fast: false
|
||||
env:
|
||||
GH_ACTION: enable
|
||||
IMAGE_NAMES: docker.io/netboxcommunity/netbox
|
||||
runs-on: ubuntu-latest
|
||||
name: Builds new NetBox Docker Images
|
||||
steps:
|
||||
- id: git-checkout
|
||||
name: Checkout
|
||||
uses: actions/checkout@v2
|
||||
- id: docker-build
|
||||
name: Build the image from '${{ matrix.docker_from }}' with '${{ matrix.build_cmd }}'
|
||||
run: ${{ matrix.build_cmd }}
|
||||
env:
|
||||
DOCKER_FROM: ${{ matrix.docker_from }}
|
||||
GH_ACTION: enable
|
||||
- id: docker-test
|
||||
name: Test the image
|
||||
run: IMAGE="${FINAL_DOCKER_TAG}" ./test.sh
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
- id: git-checkout
|
||||
name: Checkout
|
||||
uses: actions/checkout@v3
|
||||
- id: qemu-setup
|
||||
name: Set up QEMU
|
||||
uses: docker/setup-qemu-action@v2
|
||||
- id: buildx-setup
|
||||
name: Set up Docker Buildx
|
||||
uses: docker/setup-buildx-action@v2
|
||||
- id: docker-build
|
||||
name: Build the image for '${{ matrix.platform }}' with '${{ matrix.build_cmd }}'
|
||||
run: ${{ matrix.build_cmd }}
|
||||
env:
|
||||
BUILDX_PLATFORM: ${{ matrix.platform }}
|
||||
BUILDX_BUILDER_NAME: ${{ steps.buildx-setup.outputs.name }}
|
||||
- id: docker-test
|
||||
name: Test the image
|
||||
run: IMAGE="${FINAL_DOCKER_TAG}" ./test.sh
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
|
136
.github/workflows/release.yml
vendored
136
.github/workflows/release.yml
vendored
@ -1,83 +1,83 @@
|
||||
---
|
||||
name: release
|
||||
|
||||
on:
|
||||
push:
|
||||
branches:
|
||||
- release
|
||||
release:
|
||||
types:
|
||||
- published
|
||||
schedule:
|
||||
- cron: '45 5 * * *'
|
||||
workflow_dispatch:
|
||||
|
||||
jobs:
|
||||
build:
|
||||
strategy:
|
||||
matrix:
|
||||
build_cmd:
|
||||
- ./build-latest.sh
|
||||
- PRERELEASE=true ./build-latest.sh
|
||||
- ./build-next.sh
|
||||
- ./build.sh develop
|
||||
- ./build-latest.sh
|
||||
- PRERELEASE=true ./build-latest.sh
|
||||
- ./build.sh feature
|
||||
- ./build.sh develop
|
||||
platform:
|
||||
- linux/amd64,linux/arm64
|
||||
fail-fast: false
|
||||
runs-on: ubuntu-latest
|
||||
name: Builds new NetBox Docker Images
|
||||
env:
|
||||
GH_ACTION: enable
|
||||
IMAGE_NAMES: docker.io/netboxcommunity/netbox quay.io/netboxcommunity/netbox ghcr.io/netbox-community/netbox
|
||||
steps:
|
||||
- id: git-checkout
|
||||
name: Checkout
|
||||
uses: actions/checkout@v2
|
||||
- id: docker-build
|
||||
name: Build the image with '${{ matrix.build_cmd }}'
|
||||
run: ${{ matrix.build_cmd }}
|
||||
env:
|
||||
GH_ACTION: enable
|
||||
- id: docker-test
|
||||
name: Test the image
|
||||
run: IMAGE="${FINAL_DOCKER_TAG}" ./test.sh
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
- id: registry-login
|
||||
name: Login to the Docker Registry
|
||||
run: |
|
||||
echo "::add-mask::$DOCKERHUB_USERNAME"
|
||||
echo "::add-mask::$DOCKERHUB_PASSWORD"
|
||||
docker login -u "$DOCKERHUB_USERNAME" --password "${DOCKERHUB_PASSWORD}" "${DOCKER_REGISTRY}"
|
||||
env:
|
||||
DOCKERHUB_USERNAME: ${{ secrets.dockerhub_username }}
|
||||
DOCKERHUB_PASSWORD: ${{ secrets.dockerhub_password }}
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
- id: registry-push
|
||||
name: Push the image
|
||||
run: ${{ matrix.build_cmd }} --push-only
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
- id: registry-logout
|
||||
name: Logout of the Docker Registry
|
||||
run: docker logout "${DOCKER_REGISTRY}"
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
|
||||
# Quay.io
|
||||
- id: quayio-docker-build
|
||||
name: Build the image with '${{ matrix.build_cmd }}'
|
||||
run: ${{ matrix.build_cmd }}
|
||||
env:
|
||||
DOCKER_REGISTRY: quay.io
|
||||
GH_ACTION: enable
|
||||
- id: quayio-registry-login
|
||||
name: Login to the Quay.io Registry
|
||||
run: |
|
||||
echo "::add-mask::$QUAYIO_USERNAME"
|
||||
echo "::add-mask::$QUAYIO_PASSWORD"
|
||||
docker login -u "$QUAYIO_USERNAME" --password "${QUAYIO_PASSWORD}" "${DOCKER_REGISTRY}"
|
||||
env:
|
||||
DOCKER_REGISTRY: quay.io
|
||||
QUAYIO_USERNAME: ${{ secrets.quayio_username }}
|
||||
QUAYIO_PASSWORD: ${{ secrets.quayio_password }}
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
- id: quayio-registry-push
|
||||
name: Push the image
|
||||
run: ${{ matrix.build_cmd }} --push-only
|
||||
env:
|
||||
DOCKER_REGISTRY: quay.io
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
- id: quayio-registry-logout
|
||||
name: Logout of the Docker Registry
|
||||
run: docker logout "${DOCKER_REGISTRY}"
|
||||
env:
|
||||
DOCKER_REGISTRY: quay.io
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
- id: source-checkout
|
||||
name: Checkout
|
||||
uses: actions/checkout@v3
|
||||
- id: set-netbox-docker-version
|
||||
name: Get Version of NetBox Docker
|
||||
run: echo "::set-output name=version::$(cat VERSION)"
|
||||
shell: bash
|
||||
- id: qemu-setup
|
||||
name: Set up QEMU
|
||||
uses: docker/setup-qemu-action@v2
|
||||
- id: buildx-setup
|
||||
name: Set up Docker Buildx
|
||||
uses: docker/setup-buildx-action@v2
|
||||
- id: docker-build
|
||||
name: Build the image with '${{ matrix.build_cmd }}'
|
||||
run: ${{ matrix.build_cmd }}
|
||||
- id: test-image
|
||||
name: Test the image
|
||||
run: IMAGE="${FINAL_DOCKER_TAG}" ./test.sh
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
# docker.io
|
||||
- id: docker-io-login
|
||||
name: Login to docker.io
|
||||
uses: docker/login-action@v2
|
||||
with:
|
||||
registry: docker.io
|
||||
username: ${{ secrets.dockerhub_username }}
|
||||
password: ${{ secrets.dockerhub_password }}
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
# quay.io
|
||||
- id: quay-io-login
|
||||
name: Login to Quay.io
|
||||
uses: docker/login-action@v2
|
||||
with:
|
||||
registry: quay.io
|
||||
username: ${{ secrets.quayio_username }}
|
||||
password: ${{ secrets.quayio_password }}
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
# ghcr.io
|
||||
- id: ghcr-io-login
|
||||
name: Login to GitHub Container Registry
|
||||
uses: docker/login-action@v2
|
||||
with:
|
||||
registry: ghcr.io
|
||||
username: ${{ github.repository_owner }}
|
||||
password: ${{ secrets.GITHUB_TOKEN }}
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
- id: build-and-push
|
||||
name: Push the image
|
||||
run: ${{ matrix.build_cmd }} --push
|
||||
if: steps.docker-build.outputs.skipped != 'true'
|
||||
env:
|
||||
BUILDX_PLATFORM: ${{ matrix.platform }}
|
||||
BUILDX_BUILDER_NAME: ${{ steps.buildx-setup.outputs.name }}
|
||||
|
3
.gitignore
vendored
3
.gitignore
vendored
@ -7,6 +7,9 @@ configuration/*
|
||||
!configuration/configuration.py
|
||||
!configuration/extra.py
|
||||
configuration/ldap/*
|
||||
!configuration/ldap/extra.py
|
||||
!configuration/ldap/ldap_config.py
|
||||
!configuration/logging.py
|
||||
!configuration/plugins.py
|
||||
prometheus.yml
|
||||
super-linter.log
|
||||
|
@ -1,3 +1,4 @@
|
||||
ignored:
|
||||
- DL3006
|
||||
- DL3018
|
||||
- DL3008
|
||||
- DL3003
|
||||
|
5
.yamllint.yaml
Normal file
5
.yamllint.yaml
Normal file
@ -0,0 +1,5 @@
|
||||
---
|
||||
|
||||
rules:
|
||||
line-length:
|
||||
max: 120
|
113
Dockerfile
113
Dockerfile
@ -1,25 +1,23 @@
|
||||
ARG FROM
|
||||
FROM ${FROM} as builder
|
||||
|
||||
RUN apk add --no-cache \
|
||||
bash \
|
||||
build-base \
|
||||
cargo \
|
||||
RUN export DEBIAN_FRONTEND=noninteractive \
|
||||
&& apt-get update -qq \
|
||||
&& apt-get upgrade \
|
||||
--yes -qq --no-install-recommends \
|
||||
&& apt-get install \
|
||||
--yes -qq --no-install-recommends \
|
||||
build-essential \
|
||||
ca-certificates \
|
||||
cyrus-sasl-dev \
|
||||
graphviz \
|
||||
jpeg-dev \
|
||||
libevent-dev \
|
||||
libffi-dev \
|
||||
libressl-dev \
|
||||
libxslt-dev \
|
||||
musl-dev \
|
||||
openldap-dev \
|
||||
postgresql-dev \
|
||||
py3-pip \
|
||||
libldap-dev \
|
||||
libpq-dev \
|
||||
libsasl2-dev \
|
||||
libssl-dev \
|
||||
python3-dev \
|
||||
&& python3 -m venv /opt/netbox/venv \
|
||||
&& /opt/netbox/venv/bin/python3 -m pip install --upgrade \
|
||||
python3-pip \
|
||||
python3-venv \
|
||||
&& python3 -m venv /opt/netbox/venv \
|
||||
&& /opt/netbox/venv/bin/python3 -m pip install --upgrade \
|
||||
pip \
|
||||
setuptools \
|
||||
wheel
|
||||
@ -37,24 +35,30 @@ RUN /opt/netbox/venv/bin/pip install \
|
||||
ARG FROM
|
||||
FROM ${FROM} as main
|
||||
|
||||
RUN apk add --no-cache \
|
||||
bash \
|
||||
RUN export DEBIAN_FRONTEND=noninteractive \
|
||||
&& apt-get update -qq \
|
||||
&& apt-get upgrade \
|
||||
--yes -qq --no-install-recommends \
|
||||
&& apt-get install \
|
||||
--yes -qq --no-install-recommends \
|
||||
ca-certificates \
|
||||
curl \
|
||||
graphviz \
|
||||
libevent \
|
||||
libffi \
|
||||
libjpeg-turbo \
|
||||
libressl \
|
||||
libxslt \
|
||||
postgresql-libs \
|
||||
libldap-common \
|
||||
libpq5 \
|
||||
openssl \
|
||||
python3 \
|
||||
py3-pip \
|
||||
ttf-ubuntu-font-family \
|
||||
unit \
|
||||
unit-python3
|
||||
|
||||
WORKDIR /opt
|
||||
python3-distutils \
|
||||
tini \
|
||||
&& curl -sL https://nginx.org/keys/nginx_signing.key \
|
||||
> /etc/apt/trusted.gpg.d/nginx.asc && \
|
||||
echo "deb https://packages.nginx.org/unit/ubuntu/ jammy unit" \
|
||||
> /etc/apt/sources.list.d/unit.list \
|
||||
&& apt-get update -qq \
|
||||
&& apt-get install \
|
||||
--yes -qq --no-install-recommends \
|
||||
unit=1.27.0-1~jammy \
|
||||
unit-python3.10=1.27.0-1~jammy \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
COPY --from=builder /opt/netbox/venv /opt/netbox/venv
|
||||
|
||||
@ -62,7 +66,9 @@ ARG NETBOX_PATH
|
||||
COPY ${NETBOX_PATH} /opt/netbox
|
||||
|
||||
COPY docker/configuration.docker.py /opt/netbox/netbox/netbox/configuration.py
|
||||
COPY docker/ldap_config.docker.py /opt/netbox/netbox/netbox/ldap_config.py
|
||||
COPY docker/docker-entrypoint.sh /opt/netbox/docker-entrypoint.sh
|
||||
COPY docker/housekeeping.sh /opt/netbox/housekeeping.sh
|
||||
COPY docker/launch-netbox.sh /opt/netbox/launch-netbox.sh
|
||||
COPY startup_scripts/ /opt/netbox/startup_scripts/
|
||||
COPY initializers/ /opt/netbox/initializers/
|
||||
@ -74,29 +80,21 @@ WORKDIR /opt/netbox/netbox
|
||||
# Must set permissions for '/opt/netbox/netbox/media' directory
|
||||
# to g+w so that pictures can be uploaded to netbox.
|
||||
RUN mkdir -p static /opt/unit/state/ /opt/unit/tmp/ \
|
||||
&& chown -R unit:root media /opt/unit/ \
|
||||
&& chmod -R g+w media /opt/unit/ \
|
||||
&& cd /opt/netbox/ && SECRET_KEY="dummy" /opt/netbox/venv/bin/python -m mkdocs build \
|
||||
--config-file /opt/netbox/mkdocs.yml --site-dir /opt/netbox/netbox/project-static/docs/ \
|
||||
&& SECRET_KEY="dummy" /opt/netbox/venv/bin/python /opt/netbox/netbox/manage.py collectstatic --no-input
|
||||
|
||||
ENTRYPOINT [ "/opt/netbox/docker-entrypoint.sh" ]
|
||||
ENV LANG=C.UTF-8 PATH=/opt/netbox/venv/bin:$PATH
|
||||
ENTRYPOINT [ "/usr/bin/tini", "--" ]
|
||||
|
||||
CMD [ "/opt/netbox/launch-netbox.sh" ]
|
||||
CMD [ "/opt/netbox/docker-entrypoint.sh", "/opt/netbox/launch-netbox.sh" ]
|
||||
|
||||
LABEL ORIGINAL_TAG="" \
|
||||
NETBOX_GIT_BRANCH="" \
|
||||
NETBOX_GIT_REF="" \
|
||||
NETBOX_GIT_URL="" \
|
||||
# See http://label-schema.org/rc1/#build-time-labels
|
||||
# Also https://microbadger.com/labels
|
||||
org.label-schema.schema-version="1.0" \
|
||||
org.label-schema.build-date="" \
|
||||
org.label-schema.name="NetBox Docker" \
|
||||
org.label-schema.description="A container based distribution of NetBox, the free and open IPAM and DCIM solution." \
|
||||
org.label-schema.vendor="The netbox-docker contributors." \
|
||||
org.label-schema.url="https://github.com/netbox-community/netbox-docker" \
|
||||
org.label-schema.usage="https://github.com/netbox-community/netbox-docker/wiki" \
|
||||
org.label-schema.vcs-url="https://github.com/netbox-community/netbox-docker.git" \
|
||||
org.label-schema.vcs-ref="" \
|
||||
org.label-schema.version="snapshot" \
|
||||
LABEL netbox.original-tag="" \
|
||||
netbox.git-branch="" \
|
||||
netbox.git-ref="" \
|
||||
netbox.git-url="" \
|
||||
# See https://github.com/opencontainers/image-spec/blob/master/annotations.md#pre-defined-annotation-keys
|
||||
org.opencontainers.image.created="" \
|
||||
org.opencontainers.image.title="NetBox Docker" \
|
||||
@ -108,17 +106,4 @@ LABEL ORIGINAL_TAG="" \
|
||||
org.opencontainers.image.documentation="https://github.com/netbox-community/netbox-docker/wiki" \
|
||||
org.opencontainers.image.source="https://github.com/netbox-community/netbox-docker.git" \
|
||||
org.opencontainers.image.revision="" \
|
||||
org.opencontainers.image.version="snapshot"
|
||||
|
||||
#####
|
||||
## LDAP specific configuration
|
||||
#####
|
||||
|
||||
FROM main as ldap
|
||||
|
||||
RUN apk add --no-cache \
|
||||
libsasl \
|
||||
libldap \
|
||||
util-linux
|
||||
|
||||
COPY docker/ldap_config.docker.py /opt/netbox/netbox/netbox/ldap_config.py
|
||||
org.opencontainers.image.version=""
|
||||
|
122
README.md
122
README.md
@ -5,49 +5,30 @@
|
||||

|
||||

|
||||

|
||||
[][netbox-docker-microbadger]
|
||||
[][netbox-docker-microbadger]
|
||||
[][netbox-docker-license]
|
||||
|
||||
[The Github repository](netbox-docker-github) houses the components needed to build NetBox as a Docker container.
|
||||
Images are built using this code and are released to [Docker Hub][netbox-dockerhub] and [Quay.io][netbox-quayio] once a day.
|
||||
[The GitHub repository](netbox-docker-github) houses the components needed to build NetBox as a container.
|
||||
Images are built regularly using the code in that repository and are pushed to [Docker Hub][netbox-dockerhub], [Quay.io][netbox-quayio] and [GitHub Container Registry][netbox-ghcr].
|
||||
|
||||
Do you have any questions?
|
||||
Before opening an issue on Github, please join the [Network To Code][ntc-slack] Slack and ask for help in our [`#netbox-docker`][netbox-docker-slack] channel.
|
||||
Before opening an issue on Github,
|
||||
please join [our Slack][netbox-docker-slack] and ask for help in the [`#netbox-docker`][netbox-docker-slack-channel] channel.
|
||||
|
||||
[github-stargazers]: https://github.com/netbox-community/netbox-docker/stargazers
|
||||
[github-release]: https://github.com/netbox-community/netbox-docker/releases
|
||||
[netbox-docker-microbadger]: https://microbadger.com/images/netboxcommunity/netbox
|
||||
[netbox-dockerhub]: https://hub.docker.com/r/netboxcommunity/netbox/
|
||||
[netbox-docker-github]: https://github.com/netbox-community/netbox-docker/
|
||||
[ntc-slack]: http://slack.networktocode.com/
|
||||
[netbox-docker-slack]: https://slack.com/app_redirect?channel=netbox-docker&team=T09LQ7E9E
|
||||
[netbox-docker-license]: https://github.com/netbox-community/netbox-docker/blob/release/LICENSE
|
||||
[netbox-quayio]: https://quay.io/repository/netboxcommunity/netbox
|
||||
|
||||
## Docker Tags
|
||||
|
||||
* `vX.Y.Z`: These are release builds, automatically built from [the corresponding releases of NetBox][netbox-releases].
|
||||
* `latest`: These are release builds, automatically built from [the `master` branch of NetBox][netbox-master].
|
||||
* `snapshot`: These are pre-release builds, automatically built from the [`develop` branch of NetBox][netbox-develop].
|
||||
* `develop-X.Y`: These are pre-release builds, automatically built from the corresponding [branch of NetBox][netbox-branches].
|
||||
|
||||
Then there is currently one extra tags for each of the above tags:
|
||||
|
||||
* `-ldap`: Contains additional dependencies and configurations for connecting NetBox to an LDAP directory.
|
||||
[Learn more about that in our wiki][netbox-docker-ldap].
|
||||
|
||||
New images are built and published automatically every ~24h.
|
||||
|
||||
[netbox-releases]: https://github.com/netbox-community/netbox/releases
|
||||
[netbox-master]: https://github.com/netbox-community/netbox/tree/master
|
||||
[netbox-develop]: https://github.com/netbox-community/netbox/tree/develop
|
||||
[netbox-branches]: https://github.com/netbox-community/netbox/branches
|
||||
[netbox-docker-ldap]: https://github.com/netbox-community/netbox-docker/wiki/LDAP
|
||||
[netbox-ghcr]: https://github.com/netbox-community/netbox-docker/pkgs/container/netbox
|
||||
[netbox-docker-github]: https://github.com/netbox-community/netbox-docker/
|
||||
[netbox-docker-slack]: https://join.slack.com/t/netdev-community/shared_invite/zt-mtts8g0n-Sm6Wutn62q_M4OdsaIycrQ
|
||||
[netbox-docker-slack-channel]: https://netdev-community.slack.com/archives/C01P0GEVBU7
|
||||
[netbox-slack-channel]: https://netdev-community.slack.com/archives/C01P0FRSXRV
|
||||
[netbox-docker-license]: https://github.com/netbox-community/netbox-docker/blob/release/LICENSE
|
||||
|
||||
## Quickstart
|
||||
|
||||
To get NetBox Docker up and running run the following commands.
|
||||
To get _NetBox Docker_ up and running run the following commands.
|
||||
There is a more complete [_Getting Started_ guide on our wiki][wiki-getting-started] which explains every step.
|
||||
|
||||
```bash
|
||||
@ -77,19 +58,70 @@ The default credentials are:
|
||||
[wiki-getting-started]: https://github.com/netbox-community/netbox-docker/wiki/Getting-Started
|
||||
[docker-reception]: https://github.com/nxt-engineering/reception
|
||||
|
||||
## Container Image Tags
|
||||
|
||||
New container images are built and published automatically every ~24h.
|
||||
|
||||
> We recommend to use either the `vX.Y.Z-a.b.c` tags or the `vX.Y-a.b.c` tags in production!
|
||||
|
||||
* `vX.Y.Z-a.b.c`, `vX.Y-a.b.c`:
|
||||
These are release builds containing _NetBox version_ `vX.Y.Z`.
|
||||
They contain the support files of _NetBox Docker version_ `a.b.c`.
|
||||
You must use _NetBox Docker version_ `a.b.c` to guarantee the compatibility.
|
||||
These images are automatically built from [the corresponding releases of NetBox][netbox-releases].
|
||||
* `latest-a.b.c`:
|
||||
These are release builds, containing the latest stable version of NetBox.
|
||||
They contain the support files of _NetBox Docker version_ `a.b.c`.
|
||||
You must use _NetBox Docker version_ `a.b.c` to guarantee the compatibility.
|
||||
These images are automatically built from [the `master` branch of NetBox][netbox-master].
|
||||
* `snapshot-a.b.c`:
|
||||
These are prerelease builds.
|
||||
They contain the support files of _NetBox Docker version_ `a.b.c`.
|
||||
You must use _NetBox Docker version_ `a.b.c` to guarantee the compatibility.
|
||||
These images are automatically built from the [`develop` branch of NetBox][netbox-develop].
|
||||
|
||||
For each of the above tag, there is an extra tag:
|
||||
|
||||
* `vX.Y.Z`, `vX.Y`:
|
||||
This is the same version as `vX.Y.Z-a.b.c` (or `vX.Y-a.b.c`, respectively).
|
||||
It always points to the latest version of _NetBox Docker_.
|
||||
* `latest`
|
||||
This is the same version as `latest-a.b.c`.
|
||||
It always points to the latest version of _NetBox Docker_.
|
||||
* `snapshot`
|
||||
This is the same version as `snapshot-a.b.c`.
|
||||
It always points to the latest version of _NetBox Docker_.
|
||||
|
||||
Then there is currently one extra tags for each of the above tags:
|
||||
|
||||
* `-ldap`:
|
||||
These container images contain additional dependencies and configuration files for connecting NetBox to an LDAP directory.
|
||||
[Learn more about that in our wiki][netbox-docker-ldap].
|
||||
|
||||
[netbox-releases]: https://github.com/netbox-community/netbox/releases
|
||||
[netbox-master]: https://github.com/netbox-community/netbox/tree/master
|
||||
[netbox-develop]: https://github.com/netbox-community/netbox/tree/develop
|
||||
[netbox-branches]: https://github.com/netbox-community/netbox/branches
|
||||
[netbox-docker-ldap]: https://github.com/netbox-community/netbox-docker/wiki/LDAP
|
||||
|
||||
## Documentation
|
||||
|
||||
Please refer [to our wiki on Github][netbox-docker-wiki] for further information on how to use this NetBox Docker image properly.
|
||||
It covers advanced topics such as using files for secrets, deployment to Kubernetes, monitoring and configuring NAPALM or LDAP.
|
||||
Please refer [to our wiki on GitHub][netbox-docker-wiki] for further information on how to use the NetBox Docker image properly.
|
||||
The wiki covers advanced topics such as using files for secrets, configuring TLS, deployment to Kubernetes, monitoring and configuring NAPALM and LDAP.
|
||||
|
||||
Our wiki is a community effort.
|
||||
Feel free to correct errors, update outdated information or provide additional guides and insights.
|
||||
|
||||
[netbox-docker-wiki]: https://github.com/netbox-community/netbox-docker/wiki/
|
||||
|
||||
## Getting Help
|
||||
|
||||
Feel free to ask questions in our [Github Community][netbox-community] or join [our Slack channel `#netbox-docker`][netbox-docker-slack] on the [Network To Code Slack][ntc-slack],
|
||||
Feel free to ask questions in our [GitHub Community][netbox-community]
|
||||
or [join our Slack][netbox-docker-slack] and ask [in our channel `#netbox-docker`][netbox-docker-slack-channel],
|
||||
which is free to use and where there are almost always people online that can help you in the Slack channel.
|
||||
|
||||
If you need help with using NetBox or developing for it or against it's API you may find the `#netbox` channel on the same Slack instance very helpful.
|
||||
If you need help with using NetBox or developing for it or against it's API
|
||||
you may find [the `#netbox` channel][netbox-slack-channel] on the same Slack instance very helpful.
|
||||
|
||||
[netbox-community]: https://github.com/netbox-community/netbox-docker/discussions
|
||||
|
||||
@ -102,22 +134,19 @@ This project relies only on *Docker* and *docker-compose* meeting these requirem
|
||||
|
||||
To check the version installed on your system run `docker --version` and `docker-compose --version`.
|
||||
|
||||
## Breaking Changes
|
||||
|
||||
From time to time it might become necessary to re-engineer the structure of this setup.
|
||||
Things like the `docker-compose.yml` file or your Kubernetes or OpenShift configurations have to be adjusted as a consequence.
|
||||
|
||||
Since November 2019 each image built from this repo contains a `org.opencontainers.image.version` label.
|
||||
(The images contained labels since April 2018, although in November 2019 the labels' names changed.)
|
||||
You can check the label of your local image by running `docker inspect netboxcommunity/netbox:v2.7.1 --format "{{json .Config.Labels}}"`.
|
||||
## Updating
|
||||
|
||||
Please read [the release notes][releases] carefully when updating to a new image version.
|
||||
Note that the version of the NetBox Docker container image must stay in sync with the code.
|
||||
|
||||
If you update for the first time, be sure [to follow our _How To Update NetBox Docker_ guide in the wiki][netbox-docker-wiki-updating].
|
||||
|
||||
[releases]: https://github.com/netbox-community/netbox-docker/releases
|
||||
[netbox-docker-wiki-updating]: https://github.com/netbox-community/netbox-docker/wiki/Updating
|
||||
|
||||
## Rebuilding the Image
|
||||
|
||||
`./build.sh` can be used to rebuild the Docker image. See `./build.sh --help` for more information.
|
||||
`./build.sh` can be used to rebuild the container image. See `./build.sh --help` for more information.
|
||||
|
||||
For more details on custom builds [consult our wiki][netbox-docker-wiki-build].
|
||||
|
||||
@ -132,8 +161,7 @@ It runs NetBox's own unit tests and ensures that all initializers work:
|
||||
IMAGE=netboxcommunity/netbox:latest ./test.sh
|
||||
```
|
||||
|
||||
## About
|
||||
## Support
|
||||
|
||||
This repository is currently maintained and funded by [nxt][nxt].
|
||||
|
||||
[nxt]: https://nxt.engineering/en/
|
||||
This repository is currently maintained by the community.
|
||||
Please consider sponsoring the maintainers of this project.
|
||||
|
@ -1,8 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
push_image_to_registry() {
|
||||
local target_tag=$1
|
||||
echo "⏫ Pushing '${target_tag}'"
|
||||
$DRY docker push "${target_tag}"
|
||||
echo "✅ Finished pushing the Docker image '${target_tag}'."
|
||||
}
|
21
build-functions/gh-functions.sh
Normal file
21
build-functions/gh-functions.sh
Normal file
@ -0,0 +1,21 @@
|
||||
#!/bin/bash
|
||||
|
||||
###
|
||||
# A regular echo, that only prints if ${GH_ACTION} is defined.
|
||||
###
|
||||
gh_echo() {
|
||||
if [ -n "${GH_ACTION}" ]; then
|
||||
echo "${@}"
|
||||
fi
|
||||
}
|
||||
|
||||
###
|
||||
# Prints the output to the file defined in ${GITHUB_ENV}.
|
||||
# Only executes if ${GH_ACTION} is defined.
|
||||
# Example Usage: gh_env "FOO_VAR=bar_value"
|
||||
###
|
||||
gh_env() {
|
||||
if [ -n "${GH_ACTION}" ]; then
|
||||
echo "${@}" >>"${GITHUB_ENV}"
|
||||
fi
|
||||
}
|
@ -3,6 +3,14 @@
|
||||
|
||||
echo "▶️ $0 $*"
|
||||
|
||||
###
|
||||
# Check for the jq library needed for parsing JSON
|
||||
###
|
||||
if ! command -v jq; then
|
||||
echo "⚠️ jq command missing from \$PATH!"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
###
|
||||
# Checking for the presence of GITHUB_OAUTH_CLIENT_ID
|
||||
# and GITHUB_OAUTH_CLIENT_SECRET
|
||||
|
@ -1,39 +0,0 @@
|
||||
#!/bin/bash
|
||||
# Builds develop, develop-* and master branches of NetBox
|
||||
|
||||
echo "▶️ $0 $*"
|
||||
|
||||
###
|
||||
# Checking for the presence of GITHUB_OAUTH_CLIENT_ID
|
||||
# and GITHUB_OAUTH_CLIENT_SECRET
|
||||
###
|
||||
if [ -n "${GITHUB_OAUTH_CLIENT_ID}" ] && [ -n "${GITHUB_OAUTH_CLIENT_SECRET}" ]; then
|
||||
echo "🗝 Performing authenticated Github API calls."
|
||||
GITHUB_OAUTH_PARAMS="client_id=${GITHUB_OAUTH_CLIENT_ID}&client_secret=${GITHUB_OAUTH_CLIENT_SECRET}"
|
||||
else
|
||||
echo "🕶 Performing unauthenticated Github API calls. This might result in lower Github rate limits!"
|
||||
GITHUB_OAUTH_PARAMS=""
|
||||
fi
|
||||
|
||||
###
|
||||
# Calling Github to get the all branches
|
||||
###
|
||||
ORIGINAL_GITHUB_REPO="${SRC_ORG-netbox-community}/${SRC_REPO-netbox}"
|
||||
GITHUB_REPO="${GITHUB_REPO-$ORIGINAL_GITHUB_REPO}"
|
||||
URL_RELEASES="https://api.github.com/repos/${GITHUB_REPO}/branches?${GITHUB_OAUTH_PARAMS}"
|
||||
|
||||
# Composing the JQ commans to extract the most recent version number
|
||||
JQ_NEXT='map(.name) | .[] | scan("^[^v].+") | match("^(develop-).*") | .string'
|
||||
|
||||
CURL="curl -sS"
|
||||
|
||||
# Querying the Github API to fetch all branches
|
||||
NEXT=$($CURL "${URL_RELEASES}" | jq -r "$JQ_NEXT")
|
||||
|
||||
if [ -n "$NEXT" ]; then
|
||||
# shellcheck disable=SC2068
|
||||
./build.sh "${NEXT}" $@
|
||||
else
|
||||
echo "No branch matching 'develop-*' found"
|
||||
echo "::set-output name=skipped::true"
|
||||
fi
|
413
build.sh
413
build.sh
@ -6,10 +6,9 @@ echo "▶️ $0 $*"
|
||||
set -e
|
||||
|
||||
if [ "${1}x" == "x" ] || [ "${1}" == "--help" ] || [ "${1}" == "-h" ]; then
|
||||
echo "Usage: ${0} <branch> [--push|--push-only]"
|
||||
echo "Usage: ${0} <branch> [--push]"
|
||||
echo " branch The branch or tag to build. Required."
|
||||
echo " --push Pushes the built Docker image to the registry."
|
||||
echo " --push-only Only pushes the Docker image to the registry, but does not build it."
|
||||
echo ""
|
||||
echo "You can use the following ENV variables to customize the build:"
|
||||
echo " SRC_ORG Which fork of netbox to use (i.e. github.com/\${SRC_ORG}/\${SRC_REPO})."
|
||||
@ -30,15 +29,10 @@ if [ "${1}x" == "x" ] || [ "${1}" == "--help" ] || [ "${1}" == "-h" ]; then
|
||||
echo " When <branch>=master: latest"
|
||||
echo " When <branch>=develop: snapshot"
|
||||
echo " Else: same as <branch>"
|
||||
echo " DOCKER_REGISTRY The Docker repository's registry (i.e. '\${DOCKER_REGISTRY}/\${DOCKER_ORG}/\${DOCKER_REPO}'')"
|
||||
echo " IMAGE_NAMES The names used for the image including the registry"
|
||||
echo " Used for tagging the image."
|
||||
echo " Default: docker.io"
|
||||
echo " DOCKER_ORG The Docker repository's organisation (i.e. '\${DOCKER_REGISTRY}/\${DOCKER_ORG}/\${DOCKER_REPO}'')"
|
||||
echo " Used for tagging the image."
|
||||
echo " Default: netboxcommunity"
|
||||
echo " DOCKER_REPO The Docker repository's name (i.e. '\${DOCKER_REGISTRY}/\${DOCKER_ORG}/\${DOCKER_REPO}'')"
|
||||
echo " Used for tagging the image."
|
||||
echo " Default: netbox"
|
||||
echo " Default: docker.io/netboxcommunity/netbox"
|
||||
echo " Example: 'docker.io/netboxcommunity/netbox quay.io/netboxcommunity/netbox'"
|
||||
echo " DOCKER_TAG The name of the tag which is applied to the image."
|
||||
echo " Useful for pushing into another registry than hub.docker.com."
|
||||
echo " Default: \${DOCKER_REGISTRY}/\${DOCKER_ORG}/\${DOCKER_REPO}:\${TAG}"
|
||||
@ -49,10 +43,25 @@ if [ "${1}x" == "x" ] || [ "${1}" == "--help" ] || [ "${1}" == "-h" ]; then
|
||||
echo " DOCKERFILE The name of Dockerfile to use."
|
||||
echo " Default: Dockerfile"
|
||||
echo " DOCKER_FROM The base image to use."
|
||||
echo " Default: 'alpine:3.13'"
|
||||
echo " DOCKER_TARGET A specific target to build."
|
||||
echo " It's currently not possible to pass multiple targets."
|
||||
echo " Default: main ldap"
|
||||
echo " Default: 'ubuntu:22.04'"
|
||||
echo " BUILDX_PLATFORMS"
|
||||
echo " Specifies the platform(s) to build the image for."
|
||||
echo " Example: 'linux/amd64,linux/arm64'"
|
||||
echo " Default: 'linux/amd64'"
|
||||
echo " BUILDX_BUILDER_NAME"
|
||||
echo " If defined, the image build will be assigned to the given builder."
|
||||
echo " If you specify this variable, make sure that the builder exists."
|
||||
echo " If this value is not defined, a new builx builder with the directory name of the"
|
||||
echo " current directory (i.e. '$(basename "${PWD}")') is created."
|
||||
echo " Example: 'clever_lovelace'"
|
||||
echo " Default: undefined"
|
||||
echo " BUILDX_REMOVE_BUILDER"
|
||||
echo " If defined (and only if BUILDX_BUILDER_NAME is undefined),"
|
||||
echo " then the buildx builder created by this script will be removed after use."
|
||||
echo " This is useful if you build NetBox Docker on an automated system that does"
|
||||
echo " not manage the builders for you."
|
||||
echo " Example: 'on'"
|
||||
echo " Default: undefined"
|
||||
echo " HTTP_PROXY The proxy to use for http requests."
|
||||
echo " Example: http://proxy.domain.tld:3128"
|
||||
echo " Default: undefined"
|
||||
@ -95,6 +104,11 @@ if [ "${1}x" == "x" ] || [ "${1}" == "--help" ] || [ "${1}" == "-h" ]; then
|
||||
fi
|
||||
fi
|
||||
|
||||
source ./build-functions/gh-functions.sh
|
||||
|
||||
IMAGE_NAMES="${IMAGE_NAMES-docker.io/netboxcommunity/netbox}"
|
||||
IFS=' ' read -ra IMAGE_NAMES <<<"${IMAGE_NAMES}"
|
||||
|
||||
###
|
||||
# Enabling dry-run mode
|
||||
###
|
||||
@ -105,6 +119,8 @@ else
|
||||
DRY="echo"
|
||||
fi
|
||||
|
||||
gh_echo "::group::⤵️ Fetching the NetBox source code"
|
||||
|
||||
###
|
||||
# Variables for fetching the NetBox source
|
||||
###
|
||||
@ -118,6 +134,12 @@ NETBOX_PATH="${NETBOX_PATH-.netbox}"
|
||||
# Fetching the NetBox source
|
||||
###
|
||||
if [ "${2}" != "--push-only" ] && [ -z "${SKIP_GIT}" ]; then
|
||||
REMOTE_EXISTS=$(git ls-remote --heads --tags "${URL}" "${NETBOX_BRANCH}" | wc -l)
|
||||
if [ "${REMOTE_EXISTS}" == "0" ]; then
|
||||
echo "❌ Remote branch '${NETBOX_BRANCH}' not found in '${URL}'; Nothing to do"
|
||||
gh_echo "::set-output name=skipped::true"
|
||||
exit 0
|
||||
fi
|
||||
echo "🌐 Checking out '${NETBOX_BRANCH}' of NetBox from the url '${URL}' into '${NETBOX_PATH}'"
|
||||
if [ ! -d "${NETBOX_PATH}" ]; then
|
||||
$DRY git clone -q --depth 10 -b "${NETBOX_BRANCH}" "${URL}" "${NETBOX_PATH}"
|
||||
@ -125,7 +147,7 @@ if [ "${2}" != "--push-only" ] && [ -z "${SKIP_GIT}" ]; then
|
||||
|
||||
(
|
||||
$DRY cd "${NETBOX_PATH}"
|
||||
|
||||
# shellcheck disable=SC2030
|
||||
if [ -n "${HTTP_PROXY}" ]; then
|
||||
git config http.proxy "${HTTP_PROXY}"
|
||||
fi
|
||||
@ -138,6 +160,9 @@ if [ "${2}" != "--push-only" ] && [ -z "${SKIP_GIT}" ]; then
|
||||
echo "✅ Checked out NetBox"
|
||||
fi
|
||||
|
||||
gh_echo "::endgroup::"
|
||||
gh_echo "::group::🧮 Calculating Values"
|
||||
|
||||
###
|
||||
# Determining the value for DOCKERFILE
|
||||
# and checking whether it exists
|
||||
@ -157,7 +182,7 @@ fi
|
||||
# Determining the value for DOCKER_FROM
|
||||
###
|
||||
if [ -z "$DOCKER_FROM" ]; then
|
||||
DOCKER_FROM="alpine:3.13"
|
||||
DOCKER_FROM="ubuntu:22.04"
|
||||
fi
|
||||
|
||||
###
|
||||
@ -165,7 +190,7 @@ fi
|
||||
###
|
||||
BUILD_DATE="$(date -u '+%Y-%m-%dT%H:%M+00:00')"
|
||||
|
||||
if [ -d ".git" ]; then
|
||||
if [ -d ".git" ] && [ -z "${SKIP_GIT}" ]; then
|
||||
GIT_REF="$(git rev-parse HEAD)"
|
||||
fi
|
||||
|
||||
@ -173,7 +198,7 @@ fi
|
||||
PROJECT_VERSION="${PROJECT_VERSION-$(sed -e 's/^[[:space:]]*//' -e 's/[[:space:]]*$//' VERSION)}"
|
||||
|
||||
# Get the Git information from the netbox directory
|
||||
if [ -d "${NETBOX_PATH}/.git" ]; then
|
||||
if [ -d "${NETBOX_PATH}/.git" ] && [ -z "${SKIP_GIT}" ]; then
|
||||
NETBOX_GIT_REF=$(
|
||||
cd "${NETBOX_PATH}"
|
||||
git rev-parse HEAD
|
||||
@ -207,178 +232,194 @@ develop)
|
||||
esac
|
||||
|
||||
###
|
||||
# Determine targets to build
|
||||
# composing the final TARGET_DOCKER_TAG
|
||||
###
|
||||
DEFAULT_DOCKER_TARGETS=("main" "ldap")
|
||||
DOCKER_TARGETS=("${DOCKER_TARGET:-"${DEFAULT_DOCKER_TARGETS[@]}"}")
|
||||
echo "🏭 Building the following targets:" "${DOCKER_TARGETS[@]}"
|
||||
TARGET_DOCKER_TAG="${DOCKER_TAG-${TAG}}"
|
||||
TARGET_DOCKER_TAG_PROJECT="${TARGET_DOCKER_TAG}-${PROJECT_VERSION}"
|
||||
|
||||
###
|
||||
# Build each target
|
||||
# composing the additional DOCKER_SHORT_TAG,
|
||||
# i.e. "v2.6.1" becomes "v2.6",
|
||||
# which is only relevant for version tags
|
||||
# Also let "latest" follow the highest version
|
||||
###
|
||||
export DOCKER_BUILDKIT=${DOCKER_BUILDKIT-1}
|
||||
for DOCKER_TARGET in "${DOCKER_TARGETS[@]}"; do
|
||||
echo "🏗 Building the target '${DOCKER_TARGET}'"
|
||||
if [[ "${TAG}" =~ ^v([0-9]+)\.([0-9]+)\.[0-9]+$ ]]; then
|
||||
MAJOR=${BASH_REMATCH[1]}
|
||||
MINOR=${BASH_REMATCH[2]}
|
||||
|
||||
###
|
||||
# composing the final TARGET_DOCKER_TAG
|
||||
###
|
||||
TARGET_DOCKER_TAG="${DOCKER_TAG-${DOCKER_REGISTRY}/${DOCKER_ORG}/${DOCKER_REPO}:${TAG}}"
|
||||
if [ "${DOCKER_TARGET}" != "main" ]; then
|
||||
TARGET_DOCKER_TAG="${TARGET_DOCKER_TAG}-${DOCKER_TARGET}"
|
||||
fi
|
||||
if [ -n "${GH_ACTION}" ]; then
|
||||
echo "FINAL_DOCKER_TAG=${TARGET_DOCKER_TAG}" >>"$GITHUB_ENV"
|
||||
echo "::set-output name=skipped::false"
|
||||
fi
|
||||
TARGET_DOCKER_SHORT_TAG="${DOCKER_SHORT_TAG-v${MAJOR}.${MINOR}}"
|
||||
TARGET_DOCKER_LATEST_TAG="latest"
|
||||
TARGET_DOCKER_SHORT_TAG_PROJECT="${TARGET_DOCKER_SHORT_TAG}-${PROJECT_VERSION}"
|
||||
TARGET_DOCKER_LATEST_TAG_PROJECT="${TARGET_DOCKER_LATEST_TAG}-${PROJECT_VERSION}"
|
||||
fi
|
||||
|
||||
###
|
||||
# composing the additional DOCKER_SHORT_TAG,
|
||||
# i.e. "v2.6.1" becomes "v2.6",
|
||||
# which is only relevant for version tags
|
||||
# Also let "latest" follow the highest version
|
||||
###
|
||||
if [[ "${TAG}" =~ ^v([0-9]+)\.([0-9]+)\.[0-9]+$ ]]; then
|
||||
MAJOR=${BASH_REMATCH[1]}
|
||||
MINOR=${BASH_REMATCH[2]}
|
||||
|
||||
TARGET_DOCKER_SHORT_TAG="${DOCKER_SHORT_TAG-${DOCKER_REGISTRY}/${DOCKER_ORG}/${DOCKER_REPO}:v${MAJOR}.${MINOR}}"
|
||||
TARGET_DOCKER_LATEST_TAG="${DOCKER_REGISTRY}/${DOCKER_ORG}/${DOCKER_REPO}:latest"
|
||||
|
||||
if [ "${DOCKER_TARGET}" != "main" ]; then
|
||||
TARGET_DOCKER_SHORT_TAG="${TARGET_DOCKER_SHORT_TAG}-${DOCKER_TARGET}"
|
||||
TARGET_DOCKER_LATEST_TAG="${TARGET_DOCKER_LATEST_TAG}-${DOCKER_TARGET}"
|
||||
fi
|
||||
fi
|
||||
|
||||
###
|
||||
# Proceeding to buils stage, except if `--push-only` is passed
|
||||
###
|
||||
if [ "${2}" != "--push-only" ]; then
|
||||
###
|
||||
# Checking if the build is necessary,
|
||||
# meaning build only if one of those values changed:
|
||||
# - Python base image digest (Label: PYTHON_BASE_DIGEST)
|
||||
# - netbox git ref (Label: NETBOX_GIT_REF)
|
||||
# - netbox-docker git ref (Label: org.label-schema.vcs-ref)
|
||||
###
|
||||
# Load information from registry (only for docker.io)
|
||||
SHOULD_BUILD="false"
|
||||
BUILD_REASON=""
|
||||
if [ -z "${GH_ACTION}" ]; then
|
||||
# Asuming non Github builds should always proceed
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} interactive"
|
||||
elif [ "$DOCKER_REGISTRY" = "docker.io" ]; then
|
||||
source ./build-functions/get-public-image-config.sh
|
||||
IFS=':' read -ra DOCKER_FROM_SPLIT <<<"${DOCKER_FROM}"
|
||||
if ! [[ ${DOCKER_FROM_SPLIT[0]} =~ .*/.* ]]; then
|
||||
# Need to use "library/..." for images the have no two part name
|
||||
DOCKER_FROM_SPLIT[0]="library/${DOCKER_FROM_SPLIT[0]}"
|
||||
fi
|
||||
PYTHON_LAST_LAYER=$(get_image_last_layer "${DOCKER_FROM_SPLIT[0]}" "${DOCKER_FROM_SPLIT[1]}")
|
||||
mapfile -t IMAGES_LAYERS_OLD < <(get_image_layers "${DOCKER_ORG}"/"${DOCKER_REPO}" "${TAG}")
|
||||
NETBOX_GIT_REF_OLD=$(get_image_label NETBOX_GIT_REF "${DOCKER_ORG}"/"${DOCKER_REPO}" "${TAG}")
|
||||
GIT_REF_OLD=$(get_image_label org.label-schema.vcs-ref "${DOCKER_ORG}"/"${DOCKER_REPO}" "${TAG}")
|
||||
|
||||
if ! printf '%s\n' "${IMAGES_LAYERS_OLD[@]}" | grep -q -P "^${PYTHON_LAST_LAYER}\$"; then
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} alpine"
|
||||
fi
|
||||
if [ "${NETBOX_GIT_REF}" != "${NETBOX_GIT_REF_OLD}" ]; then
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} netbox"
|
||||
fi
|
||||
if [ "${GIT_REF}" != "${GIT_REF_OLD}" ]; then
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} netbox-docker"
|
||||
fi
|
||||
else
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} no-check"
|
||||
fi
|
||||
###
|
||||
# Composing all arguments for `docker build`
|
||||
###
|
||||
DOCKER_BUILD_ARGS=(
|
||||
--pull
|
||||
--target "${DOCKER_TARGET}"
|
||||
-f "${DOCKERFILE}"
|
||||
-t "${TARGET_DOCKER_TAG}"
|
||||
)
|
||||
if [ -n "${TARGET_DOCKER_SHORT_TAG}" ]; then
|
||||
DOCKER_BUILD_ARGS+=(-t "${TARGET_DOCKER_SHORT_TAG}")
|
||||
DOCKER_BUILD_ARGS+=(-t "${TARGET_DOCKER_LATEST_TAG}")
|
||||
fi
|
||||
|
||||
# --label
|
||||
DOCKER_BUILD_ARGS+=(
|
||||
--label "ORIGINAL_TAG=${TARGET_DOCKER_TAG}"
|
||||
|
||||
--label "org.label-schema.build-date=${BUILD_DATE}"
|
||||
--label "org.opencontainers.image.created=${BUILD_DATE}"
|
||||
|
||||
--label "org.label-schema.version=${PROJECT_VERSION}"
|
||||
--label "org.opencontainers.image.version=${PROJECT_VERSION}"
|
||||
)
|
||||
if [ -d ".git" ]; then
|
||||
DOCKER_BUILD_ARGS+=(
|
||||
--label "org.label-schema.vcs-ref=${GIT_REF}"
|
||||
--label "org.opencontainers.image.revision=${GIT_REF}"
|
||||
)
|
||||
fi
|
||||
if [ -d "${NETBOX_PATH}/.git" ]; then
|
||||
DOCKER_BUILD_ARGS+=(
|
||||
--label "NETBOX_GIT_BRANCH=${NETBOX_GIT_BRANCH}"
|
||||
--label "NETBOX_GIT_REF=${NETBOX_GIT_REF}"
|
||||
--label "NETBOX_GIT_URL=${NETBOX_GIT_URL}"
|
||||
)
|
||||
fi
|
||||
if [ -n "${BUILD_REASON}" ]; then
|
||||
BUILD_REASON=$(sed -e 's/^[[:space:]]*//' -e 's/[[:space:]]*$//' <<<"$BUILD_REASON")
|
||||
DOCKER_BUILD_ARGS+=(--label "BUILD_REASON=${BUILD_REASON}")
|
||||
fi
|
||||
|
||||
# --build-arg
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "NETBOX_PATH=${NETBOX_PATH}")
|
||||
|
||||
if [ -n "${DOCKER_FROM}" ]; then
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "FROM=${DOCKER_FROM}")
|
||||
fi
|
||||
if [ -n "${HTTP_PROXY}" ]; then
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "http_proxy=${HTTP_PROXY}")
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "https_proxy=${HTTPS_PROXY}")
|
||||
fi
|
||||
if [ -n "${NO_PROXY}" ]; then
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "no_proxy=${NO_PROXY}")
|
||||
fi
|
||||
|
||||
###
|
||||
# Building the docker image
|
||||
###
|
||||
if [ "${SHOULD_BUILD}" == "true" ]; then
|
||||
echo "🐳 Building the Docker image '${TARGET_DOCKER_TAG}'."
|
||||
echo " Build reason set to: ${BUILD_REASON}"
|
||||
$DRY docker build "${DOCKER_BUILD_ARGS[@]}" .
|
||||
echo "✅ Finished building the Docker images '${TARGET_DOCKER_TAG}'"
|
||||
echo "🔎 Inspecting labels on '${TARGET_DOCKER_TAG}'"
|
||||
$DRY docker inspect "${TARGET_DOCKER_TAG}" --format "{{json .Config.Labels}}"
|
||||
else
|
||||
echo "Build skipped because sources didn't change"
|
||||
echo "::set-output name=skipped::true"
|
||||
fi
|
||||
fi
|
||||
|
||||
###
|
||||
# Pushing the docker images if either `--push` or `--push-only` are passed
|
||||
###
|
||||
if [ "${2}" == "--push" ] || [ "${2}" == "--push-only" ]; then
|
||||
source ./build-functions/docker-functions.sh
|
||||
push_image_to_registry "${TARGET_DOCKER_TAG}"
|
||||
|
||||
if [ -n "${TARGET_DOCKER_SHORT_TAG}" ]; then
|
||||
push_image_to_registry "${TARGET_DOCKER_SHORT_TAG}"
|
||||
push_image_to_registry "${TARGET_DOCKER_LATEST_TAG}"
|
||||
fi
|
||||
fi
|
||||
IMAGE_NAME_TAGS=()
|
||||
for IMAGE_NAME in "${IMAGE_NAMES[@]}"; do
|
||||
IMAGE_NAME_TAGS+=("${IMAGE_NAME}:${TARGET_DOCKER_TAG}")
|
||||
IMAGE_NAME_TAGS+=("${IMAGE_NAME}:${TARGET_DOCKER_TAG_PROJECT}")
|
||||
done
|
||||
if [ -n "${TARGET_DOCKER_SHORT_TAG}" ]; then
|
||||
for IMAGE_NAME in "${IMAGE_NAMES[@]}"; do
|
||||
IMAGE_NAME_TAGS+=("${IMAGE_NAME}:${TARGET_DOCKER_SHORT_TAG}")
|
||||
IMAGE_NAME_TAGS+=("${IMAGE_NAME}:${TARGET_DOCKER_SHORT_TAG_PROJECT}")
|
||||
IMAGE_NAME_TAGS+=("${IMAGE_NAME}:${TARGET_DOCKER_LATEST_TAG}")
|
||||
IMAGE_NAME_TAGS+=("${IMAGE_NAME}:${TARGET_DOCKER_LATEST_TAG_PROJECT}")
|
||||
done
|
||||
fi
|
||||
|
||||
gh_env "FINAL_DOCKER_TAG=${IMAGE_NAME_TAGS[0]}"
|
||||
|
||||
###
|
||||
# Checking if the build is necessary,
|
||||
# meaning build only if one of those values changed:
|
||||
# - base image digest
|
||||
# - netbox git ref (Label: netbox.git-ref)
|
||||
# - netbox-docker git ref (Label: org.opencontainers.image.revision)
|
||||
###
|
||||
# Load information from registry (only for docker.io)
|
||||
SHOULD_BUILD="false"
|
||||
BUILD_REASON=""
|
||||
if [ -z "${GH_ACTION}" ]; then
|
||||
# Asuming non Github builds should always proceed
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} interactive"
|
||||
elif [[ "${IMAGE_NAME_TAGS[0]}" = docker.io* ]]; then
|
||||
source ./build-functions/get-public-image-config.sh
|
||||
IFS=':' read -ra DOCKER_FROM_SPLIT <<<"${DOCKER_FROM}"
|
||||
if ! [[ ${DOCKER_FROM_SPLIT[0]} =~ .*/.* ]]; then
|
||||
# Need to use "library/..." for images the have no two part name
|
||||
DOCKER_FROM_SPLIT[0]="library/${DOCKER_FROM_SPLIT[0]}"
|
||||
fi
|
||||
IFS='/' read -ra ORG_REPO <<<"${IMAGE_NAMES[0]}"
|
||||
echo "Checking labels for '${ORG_REPO[1]}' and '${ORG_REPO[2]}'"
|
||||
BASE_LAST_LAYER=$(get_image_last_layer "${DOCKER_FROM_SPLIT[0]}" "${DOCKER_FROM_SPLIT[1]}")
|
||||
mapfile -t IMAGES_LAYERS_OLD < <(get_image_layers "${ORG_REPO[1]}"/"${ORG_REPO[2]}" "${TAG}")
|
||||
NETBOX_GIT_REF_OLD=$(get_image_label netbox.git-ref "${ORG_REPO[1]}"/"${ORG_REPO[2]}" "${TAG}")
|
||||
GIT_REF_OLD=$(get_image_label org.opencontainers.image.revision "${ORG_REPO[1]}"/"${ORG_REPO[2]}" "${TAG}")
|
||||
|
||||
if ! printf '%s\n' "${IMAGES_LAYERS_OLD[@]}" | grep -q -P "^${BASE_LAST_LAYER}\$"; then
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} debian"
|
||||
fi
|
||||
if [ "${NETBOX_GIT_REF}" != "${NETBOX_GIT_REF_OLD}" ]; then
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} netbox"
|
||||
fi
|
||||
if [ "${GIT_REF}" != "${GIT_REF_OLD}" ]; then
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} netbox-docker"
|
||||
fi
|
||||
else
|
||||
SHOULD_BUILD="true"
|
||||
BUILD_REASON="${BUILD_REASON} no-check"
|
||||
fi
|
||||
|
||||
if [ "${SHOULD_BUILD}" != "true" ]; then
|
||||
echo "Build skipped because sources didn't change"
|
||||
echo "::set-output name=skipped::true"
|
||||
exit 0 # Nothing to do -> exit
|
||||
else
|
||||
gh_echo "::set-output name=skipped::false"
|
||||
fi
|
||||
gh_echo "::endgroup::"
|
||||
|
||||
###
|
||||
# Build the image
|
||||
###
|
||||
gh_echo "::group::🏗 Building the image"
|
||||
###
|
||||
# Composing all arguments for `docker build`
|
||||
###
|
||||
DOCKER_BUILD_ARGS=(
|
||||
--pull
|
||||
--target main
|
||||
-f "${DOCKERFILE}"
|
||||
)
|
||||
for IMAGE_NAME in "${IMAGE_NAME_TAGS[@]}"; do
|
||||
DOCKER_BUILD_ARGS+=(-t "${IMAGE_NAME}")
|
||||
done
|
||||
|
||||
# --label
|
||||
DOCKER_BUILD_ARGS+=(
|
||||
--label "netbox.original-tag=${TARGET_DOCKER_TAG_PROJECT}"
|
||||
--label "org.opencontainers.image.created=${BUILD_DATE}"
|
||||
--label "org.opencontainers.image.version=${PROJECT_VERSION}"
|
||||
)
|
||||
if [ -d ".git" ] && [ -z "${SKIP_GIT}" ]; then
|
||||
DOCKER_BUILD_ARGS+=(
|
||||
--label "org.opencontainers.image.revision=${GIT_REF}"
|
||||
)
|
||||
fi
|
||||
if [ -d "${NETBOX_PATH}/.git" ] && [ -z "${SKIP_GIT}" ]; then
|
||||
DOCKER_BUILD_ARGS+=(
|
||||
--label "netbox.git-branch=${NETBOX_GIT_BRANCH}"
|
||||
--label "netbox.git-ref=${NETBOX_GIT_REF}"
|
||||
--label "netbox.git-url=${NETBOX_GIT_URL}"
|
||||
)
|
||||
fi
|
||||
if [ -n "${BUILD_REASON}" ]; then
|
||||
BUILD_REASON=$(sed -e 's/^[[:space:]]*//' -e 's/[[:space:]]*$//' <<<"$BUILD_REASON")
|
||||
DOCKER_BUILD_ARGS+=(--label "netbox.build-reason=${BUILD_REASON}")
|
||||
fi
|
||||
|
||||
# --build-arg
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "NETBOX_PATH=${NETBOX_PATH}")
|
||||
|
||||
if [ -n "${DOCKER_FROM}" ]; then
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "FROM=${DOCKER_FROM}")
|
||||
fi
|
||||
# shellcheck disable=SC2031
|
||||
if [ -n "${HTTP_PROXY}" ]; then
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "http_proxy=${HTTP_PROXY}")
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "https_proxy=${HTTPS_PROXY}")
|
||||
fi
|
||||
if [ -n "${NO_PROXY}" ]; then
|
||||
DOCKER_BUILD_ARGS+=(--build-arg "no_proxy=${NO_PROXY}")
|
||||
fi
|
||||
|
||||
DOCKER_BUILD_ARGS+=(--platform "${BUILDX_PLATFORM-linux/amd64}")
|
||||
if [ "${2}" == "--push" ]; then
|
||||
# output type=docker does not work with pushing
|
||||
DOCKER_BUILD_ARGS+=(
|
||||
--output=type=image
|
||||
--push
|
||||
)
|
||||
else
|
||||
DOCKER_BUILD_ARGS+=(
|
||||
--output=type=docker
|
||||
)
|
||||
fi
|
||||
|
||||
###
|
||||
# Building the docker image
|
||||
###
|
||||
if [ -z "${BUILDX_BUILDER_NAME}" ]; then
|
||||
BUILDX_BUILDER_NAME="$(basename "${PWD}")"
|
||||
fi
|
||||
if ! docker buildx ls | grep --quiet --word-regexp "${BUILDX_BUILDER_NAME}"; then
|
||||
echo "👷 Creating new Buildx Builder '${BUILDX_BUILDER_NAME}'"
|
||||
$DRY docker buildx create --name "${BUILDX_BUILDER_NAME}"
|
||||
BUILDX_BUILDER_CREATED="yes"
|
||||
fi
|
||||
|
||||
echo "🐳 Building the Docker image '${TARGET_DOCKER_TAG_PROJECT}'."
|
||||
echo " Build reason set to: ${BUILD_REASON}"
|
||||
$DRY docker buildx \
|
||||
--builder "${BUILDX_BUILDER_NAME}" \
|
||||
build \
|
||||
"${DOCKER_BUILD_ARGS[@]}" \
|
||||
.
|
||||
echo "✅ Finished building the Docker images"
|
||||
gh_echo "::endgroup::" # End group for Build
|
||||
|
||||
gh_echo "::group::🏗 Image Labels"
|
||||
echo "🔎 Inspecting labels on '${IMAGE_NAME_TAGS[0]}'"
|
||||
$DRY docker inspect "${IMAGE_NAME_TAGS[0]}" --format "{{json .Config.Labels}}" | jq
|
||||
gh_echo "::endgroup::"
|
||||
|
||||
gh_echo "::group::🏗 Clean up"
|
||||
if [ -n "${BUILDX_REMOVE_BUILDER}" ] && [ "${BUILDX_BUILDER_CREATED}" == "yes" ]; then
|
||||
echo "👷 Removing Buildx Builder '${BUILDX_BUILDER_NAME}'"
|
||||
$DRY docker buildx rm "${BUILDX_BUILDER_NAME}"
|
||||
fi
|
||||
gh_echo "::endgroup::"
|
||||
|
@ -48,6 +48,8 @@ DATABASE = {
|
||||
# Database connection SSLMODE
|
||||
'CONN_MAX_AGE': int(environ.get('DB_CONN_MAX_AGE', '300')),
|
||||
# Max database connection age
|
||||
'DISABLE_SERVER_SIDE_CURSORS': environ.get('DB_DISABLE_SERVER_SIDE_CURSORS', 'False').lower() == 'true',
|
||||
# Disable the use of server-side cursors transaction pooling
|
||||
}
|
||||
|
||||
# Redis database settings. Redis is used for caching and for queuing background tasks such as webhook events. A separate
|
||||
@ -60,6 +62,7 @@ REDIS = {
|
||||
'PASSWORD': _read_secret('redis_password', environ.get('REDIS_PASSWORD', '')),
|
||||
'DATABASE': int(environ.get('REDIS_DATABASE', 0)),
|
||||
'SSL': environ.get('REDIS_SSL', 'False').lower() == 'true',
|
||||
'INSECURE_SKIP_TLS_VERIFY': environ.get('REDIS_INSECURE_SKIP_TLS_VERIFY', 'False').lower() == 'true',
|
||||
},
|
||||
'caching': {
|
||||
'HOST': environ.get('REDIS_CACHE_HOST', environ.get('REDIS_HOST', 'localhost')),
|
||||
@ -67,6 +70,7 @@ REDIS = {
|
||||
'PASSWORD': _read_secret('redis_cache_password', environ.get('REDIS_CACHE_PASSWORD', environ.get('REDIS_PASSWORD', ''))),
|
||||
'DATABASE': int(environ.get('REDIS_CACHE_DATABASE', 1)),
|
||||
'SSL': environ.get('REDIS_CACHE_SSL', environ.get('REDIS_SSL', 'False')).lower() == 'true',
|
||||
'INSECURE_SKIP_TLS_VERIFY': environ.get('REDIS_CACHE_INSECURE_SKIP_TLS_VERIFY', environ.get('REDIS_INSECURE_SKIP_TLS_VERIFY', 'False')).lower() == 'true',
|
||||
},
|
||||
}
|
||||
|
||||
@ -106,9 +110,6 @@ BANNER_LOGIN = environ.get('BANNER_LOGIN', '')
|
||||
# BASE_PATH = 'netbox/'
|
||||
BASE_PATH = environ.get('BASE_PATH', '')
|
||||
|
||||
# Cache timeout in seconds. Set to 0 to dissable caching. Defaults to 900 (15 minutes)
|
||||
CACHE_TIMEOUT = int(environ.get('CACHE_TIMEOUT', 900))
|
||||
|
||||
# Maximum number of days to retain logged changes. Set to 0 to retain changes indefinitely. (Default: 90)
|
||||
CHANGELOG_RETENTION = int(environ.get('CHANGELOG_RETENTION', 90))
|
||||
|
||||
@ -119,6 +120,11 @@ CORS_ORIGIN_ALLOW_ALL = environ.get('CORS_ORIGIN_ALLOW_ALL', 'False').lower() ==
|
||||
CORS_ORIGIN_WHITELIST = list(filter(None, environ.get('CORS_ORIGIN_WHITELIST', 'https://localhost').split(' ')))
|
||||
CORS_ORIGIN_REGEX_WHITELIST = [re.compile(r) for r in list(filter(None, environ.get('CORS_ORIGIN_REGEX_WHITELIST', '').split(' ')))]
|
||||
|
||||
# Cross-Site-Request-Forgery-Attack settings. If Netbox is sitting behind a reverse proxy, you might need to set the CSRF_TRUSTED_ORIGINS flag.
|
||||
# Django 4.0 requires to specify the URL Scheme in this setting. An example environment variable could be specified like:
|
||||
# CSRF_TRUSTED_ORIGINS=https://demo.netbox.dev http://demo.netbox.dev
|
||||
CSRF_TRUSTED_ORIGINS = list(filter(None, environ.get('CSRF_TRUSTED_ORIGINS', '').split(' ')))
|
||||
|
||||
# Set to True to enable server debugging. WARNING: Debugging introduces a substantial performance penalty and may reveal
|
||||
# sensitive information about your installation. Only enable debugging while performing testing. Never enable debugging
|
||||
# on a production system.
|
||||
@ -146,6 +152,9 @@ ENFORCE_GLOBAL_UNIQUE = environ.get('ENFORCE_GLOBAL_UNIQUE', 'False').lower() ==
|
||||
# by anonymous users. List models in the form `<app>.<model>`. Add '*' to this list to exempt all models.
|
||||
EXEMPT_VIEW_PERMISSIONS = list(filter(None, environ.get('EXEMPT_VIEW_PERMISSIONS', '').split(' ')))
|
||||
|
||||
# Enable GraphQL API.
|
||||
GRAPHQL_ENABLED = environ.get('GRAPHQL_ENABLED', 'True').lower() == 'true'
|
||||
|
||||
# Enable custom logging. Please see the Django documentation for detailed guidance on configuring custom logs:
|
||||
# https://docs.djangoproject.com/en/stable/topics/logging/
|
||||
LOGGING = {}
|
||||
@ -161,6 +170,9 @@ LOGIN_TIMEOUT = int(environ.get('LOGIN_TIMEOUT', 1209600))
|
||||
# Setting this to True will display a "maintenance mode" banner at the top of every page.
|
||||
MAINTENANCE_MODE = environ.get('MAINTENANCE_MODE', 'False').lower() == 'true'
|
||||
|
||||
# Maps provider
|
||||
MAPS_URL = environ.get('MAPS_URL', None)
|
||||
|
||||
# An API consumer can request an arbitrary number of objects =by appending the "limit" parameter to the URL (e.g.
|
||||
# "?limit=1000"). This setting defines the maximum limit. Setting it to 0 or None will allow an API consumer to request
|
||||
# all objects by specifying "?limit=0".
|
||||
@ -210,9 +222,6 @@ REMOTE_AUTH_HEADER = environ.get('REMOTE_AUTH_HEADER', 'HTTP_REMOTE_USER')
|
||||
REMOTE_AUTH_AUTO_CREATE_USER = environ.get('REMOTE_AUTH_AUTO_CREATE_USER', 'True').lower() == 'true'
|
||||
REMOTE_AUTH_DEFAULT_GROUPS = list(filter(None, environ.get('REMOTE_AUTH_DEFAULT_GROUPS', '').split(' ')))
|
||||
|
||||
# This determines how often the GitHub API is called to check the latest release of NetBox. Must be at least 1 hour.
|
||||
RELEASE_CHECK_TIMEOUT = int(environ.get('RELEASE_CHECK_TIMEOUT', 24 * 3600))
|
||||
|
||||
# This repository is used to check whether there is a new release of NetBox available. Set to None to disable the
|
||||
# version check or use the URL below to check for release in the official NetBox repository.
|
||||
# https://api.github.com/repos/netbox-community/netbox/releases
|
||||
|
28
configuration/ldap/extra.py
Normal file
28
configuration/ldap/extra.py
Normal file
@ -0,0 +1,28 @@
|
||||
####
|
||||
## This file contains extra configuration options that can't be configured
|
||||
## directly through environment variables.
|
||||
## All vairables set here overwrite any existing found in ldap_config.py
|
||||
####
|
||||
|
||||
# # This Python script inherits all the imports from ldap_config.py
|
||||
# from django_auth_ldap.config import LDAPGroupQuery # Imported since not in ldap_config.py
|
||||
|
||||
# # Sets a base requirement of membetship to netbox-user-ro, netbox-user-rw, or netbox-user-admin.
|
||||
# AUTH_LDAP_REQUIRE_GROUP = (
|
||||
# LDAPGroupQuery("cn=netbox-user-ro,ou=groups,dc=example,dc=com")
|
||||
# | LDAPGroupQuery("cn=netbox-user-rw,ou=groups,dc=example,dc=com")
|
||||
# | LDAPGroupQuery("cn=netbox-user-admin,ou=groups,dc=example,dc=com")
|
||||
# )
|
||||
|
||||
# # Sets LDAP Flag groups variables with example.
|
||||
# AUTH_LDAP_USER_FLAGS_BY_GROUP = {
|
||||
# "is_staff": (
|
||||
# LDAPGroupQuery("cn=netbox-user-ro,ou=groups,dc=example,dc=com")
|
||||
# | LDAPGroupQuery("cn=netbox-user-rw,ou=groups,dc=example,dc=com")
|
||||
# | LDAPGroupQuery("cn=netbox-user-admin,ou=groups,dc=example,dc=com")
|
||||
# ),
|
||||
# "is_superuser": "cn=netbox-user-admin,ou=groups,dc=example,dc=com",
|
||||
# }
|
||||
|
||||
# # Sets LDAP Mirror groups variables with example groups
|
||||
# AUTH_LDAP_MIRROR_GROUPS = ["netbox-user-ro", "netbox-user-rw", "netbox-user-admin"]
|
55
configuration/logging.py
Normal file
55
configuration/logging.py
Normal file
@ -0,0 +1,55 @@
|
||||
# # Remove first comment(#) on each line to implement this working logging example.
|
||||
# # Add LOGLEVEL environment variable to netbox if you use this example & want a different log level.
|
||||
# from os import environ
|
||||
|
||||
# # Set LOGLEVEL in netbox.env or docker-compose.overide.yml to override a logging level of INFO.
|
||||
# LOGLEVEL = environ.get('LOGLEVEL', 'INFO')
|
||||
|
||||
# LOGGING = {
|
||||
|
||||
# 'version': 1,
|
||||
# 'disable_existing_loggers': False,
|
||||
# 'formatters': {
|
||||
# 'verbose': {
|
||||
# 'format': '{levelname} {asctime} {module} {process:d} {thread:d} {message}',
|
||||
# 'style': '{',
|
||||
# },
|
||||
# 'simple': {
|
||||
# 'format': '{levelname} {message}',
|
||||
# 'style': '{',
|
||||
# },
|
||||
# },
|
||||
# 'filters': {
|
||||
# 'require_debug_false': {
|
||||
# '()': 'django.utils.log.RequireDebugFalse',
|
||||
# },
|
||||
# },
|
||||
# 'handlers': {
|
||||
# 'console': {
|
||||
# 'level': LOGLEVEL,
|
||||
# 'filters': ['require_debug_false'],
|
||||
# 'class': 'logging.StreamHandler',
|
||||
# 'formatter': 'simple'
|
||||
# },
|
||||
# 'mail_admins': {
|
||||
# 'level': 'ERROR',
|
||||
# 'class': 'django.utils.log.AdminEmailHandler',
|
||||
# 'filters': ['require_debug_false']
|
||||
# }
|
||||
# },
|
||||
# 'loggers': {
|
||||
# 'django': {
|
||||
# 'handlers': ['console'],
|
||||
# 'propagate': True,
|
||||
# },
|
||||
# 'django.request': {
|
||||
# 'handlers': ['mail_admins'],
|
||||
# 'level': 'ERROR',
|
||||
# 'propagate': False,
|
||||
# },
|
||||
# 'django_auth_ldap': {
|
||||
# 'handlers': ['console',],
|
||||
# 'level': LOGLEVEL,
|
||||
# }
|
||||
# }
|
||||
# }
|
13
configuration/plugins.py
Normal file
13
configuration/plugins.py
Normal file
@ -0,0 +1,13 @@
|
||||
# Add your plugins and plugin settings here.
|
||||
# Of course uncomment this file out.
|
||||
|
||||
# To learn how to build images with your required plugins
|
||||
# See https://github.com/netbox-community/netbox-docker/wiki/Using-Netbox-Plugins
|
||||
|
||||
# PLUGINS = ["netbox_bgp"]
|
||||
|
||||
# PLUGINS_CONFIG = {
|
||||
# "netbox_bgp": {
|
||||
# ADD YOUR SETTINGS HERE
|
||||
# }
|
||||
# }
|
5
docker-compose.override.yml.example
Normal file
5
docker-compose.override.yml.example
Normal file
@ -0,0 +1,5 @@
|
||||
version: '3.4'
|
||||
services:
|
||||
netbox:
|
||||
ports:
|
||||
- 8000:8080
|
@ -9,7 +9,7 @@ services:
|
||||
env_file: env/netbox.env
|
||||
environment:
|
||||
SKIP_STARTUP_SCRIPTS: ${SKIP_STARTUP_SCRIPTS-false}
|
||||
user: '101'
|
||||
user: 'unit:root'
|
||||
volumes:
|
||||
- ./startup_scripts:/opt/netbox/startup_scripts:z,ro
|
||||
- ./${INITIALIZERS_DIR-initializers}:/opt/netbox/initializers:z,ro
|
||||
@ -17,20 +17,18 @@ services:
|
||||
- ./reports:/etc/netbox/reports:z,ro
|
||||
- ./scripts:/etc/netbox/scripts:z,ro
|
||||
- netbox-media-files:/opt/netbox/netbox/media:z
|
||||
ports:
|
||||
- 8080
|
||||
postgres:
|
||||
image: postgres:12-alpine
|
||||
image: postgres:14-alpine
|
||||
env_file: env/postgres.env
|
||||
redis:
|
||||
image: redis:6-alpine
|
||||
image: redis:7-alpine
|
||||
command:
|
||||
- sh
|
||||
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
||||
- redis-server --appendonly yes --requirepass $$REDIS_PASSWORD ## $$ because of docker-compose
|
||||
env_file: env/redis.env
|
||||
redis-cache:
|
||||
image: redis:6-alpine
|
||||
image: redis:7-alpine
|
||||
command:
|
||||
- sh
|
||||
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
||||
|
@ -1,14 +1,14 @@
|
||||
version: '3.4'
|
||||
services:
|
||||
netbox: &netbox
|
||||
image: netboxcommunity/netbox:${VERSION-latest}
|
||||
image: netboxcommunity/netbox:${VERSION-v3.2-2.1.0}
|
||||
depends_on:
|
||||
- postgres
|
||||
- redis
|
||||
- redis-cache
|
||||
- netbox-worker
|
||||
env_file: env/netbox.env
|
||||
user: '101'
|
||||
user: 'unit:root'
|
||||
volumes:
|
||||
- ./startup_scripts:/opt/netbox/startup_scripts:z,ro
|
||||
- ./initializers:/opt/netbox/initializers:z,ro
|
||||
@ -16,29 +16,33 @@ services:
|
||||
- ./reports:/etc/netbox/reports:z,ro
|
||||
- ./scripts:/etc/netbox/scripts:z,ro
|
||||
- netbox-media-files:/opt/netbox/netbox/media:z
|
||||
ports:
|
||||
- "8080"
|
||||
netbox-worker:
|
||||
<<: *netbox
|
||||
depends_on:
|
||||
- redis
|
||||
entrypoint:
|
||||
- postgres
|
||||
command:
|
||||
- /opt/netbox/venv/bin/python
|
||||
- /opt/netbox/netbox/manage.py
|
||||
command:
|
||||
- rqworker
|
||||
ports: []
|
||||
netbox-housekeeping:
|
||||
<<: *netbox
|
||||
depends_on:
|
||||
- redis
|
||||
- postgres
|
||||
command:
|
||||
- /opt/netbox/housekeeping.sh
|
||||
|
||||
# postgres
|
||||
postgres:
|
||||
image: postgres:12-alpine
|
||||
image: postgres:14-alpine
|
||||
env_file: env/postgres.env
|
||||
volumes:
|
||||
- netbox-postgres-data:/var/lib/postgresql/data
|
||||
|
||||
# redis
|
||||
redis:
|
||||
image: redis:6-alpine
|
||||
image: redis:7-alpine
|
||||
command:
|
||||
- sh
|
||||
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
||||
@ -47,7 +51,7 @@ services:
|
||||
volumes:
|
||||
- netbox-redis-data:/data
|
||||
redis-cache:
|
||||
image: redis:6-alpine
|
||||
image: redis:7-alpine
|
||||
command:
|
||||
- sh
|
||||
- -c # this is to evaluate the $REDIS_PASSWORD from the env
|
||||
|
@ -45,6 +45,9 @@ def read_configurations(config_module, config_dir, main_config):
|
||||
if not f.name.endswith(".py"):
|
||||
continue
|
||||
|
||||
if f.name == f"{main_config}.py":
|
||||
continue
|
||||
|
||||
if f.name == f"{config_dir}.py":
|
||||
continue
|
||||
|
||||
@ -79,3 +82,10 @@ def __getattr__(name):
|
||||
except:
|
||||
pass
|
||||
raise AttributeError
|
||||
|
||||
|
||||
def __dir__():
|
||||
names = []
|
||||
for config in _loaded_configurations:
|
||||
names.extend(config.__dir__())
|
||||
return names
|
||||
|
@ -15,7 +15,19 @@ source /opt/netbox/venv/bin/activate
|
||||
DB_WAIT_TIMEOUT=${DB_WAIT_TIMEOUT-3}
|
||||
MAX_DB_WAIT_TIME=${MAX_DB_WAIT_TIME-30}
|
||||
CUR_DB_WAIT_TIME=0
|
||||
while ! ./manage.py migrate 2>&1 && [ "${CUR_DB_WAIT_TIME}" -lt "${MAX_DB_WAIT_TIME}" ]; do
|
||||
while [ "${CUR_DB_WAIT_TIME}" -lt "${MAX_DB_WAIT_TIME}" ]; do
|
||||
# Read and truncate connection error tracebacks to last line by default
|
||||
exec {psfd}< <(./manage.py showmigrations 2>&1)
|
||||
read -rd '' DB_ERR <&$psfd || :
|
||||
exec {psfd}<&-
|
||||
wait $! && break
|
||||
if [ -n "$DB_WAIT_DEBUG" ]; then
|
||||
echo "$DB_ERR"
|
||||
else
|
||||
readarray -tn 0 DB_ERR_LINES <<<"$DB_ERR"
|
||||
echo "${DB_ERR_LINES[@]: -1}"
|
||||
echo "[ Use DB_WAIT_DEBUG=1 in netbox.env to print full traceback for errors here ]"
|
||||
fi
|
||||
echo "⏳ Waiting on DB... (${CUR_DB_WAIT_TIME}s / ${MAX_DB_WAIT_TIME}s)"
|
||||
sleep "${DB_WAIT_TIMEOUT}"
|
||||
CUR_DB_WAIT_TIME=$((CUR_DB_WAIT_TIME + DB_WAIT_TIMEOUT))
|
||||
@ -24,6 +36,17 @@ if [ "${CUR_DB_WAIT_TIME}" -ge "${MAX_DB_WAIT_TIME}" ]; then
|
||||
echo "❌ Waited ${MAX_DB_WAIT_TIME}s or more for the DB to become ready."
|
||||
exit 1
|
||||
fi
|
||||
# Check if update is needed
|
||||
if ! ./manage.py migrate --check >/dev/null 2>&1; then
|
||||
echo "⚙️ Applying database migrations"
|
||||
./manage.py migrate --no-input
|
||||
echo "⚙️ Running trace_paths"
|
||||
./manage.py trace_paths --no-input
|
||||
echo "⚙️ Removing stale content types"
|
||||
./manage.py remove_stale_contenttypes --no-input
|
||||
echo "⚙️ Removing expired user sessions"
|
||||
./manage.py clearsessions
|
||||
fi
|
||||
|
||||
# Create Superuser if required
|
||||
if [ "$SKIP_SUPERUSER" == "true" ]; then
|
||||
@ -68,6 +91,4 @@ echo "✅ Initialisation is done."
|
||||
|
||||
# Launch whatever is passed by docker
|
||||
# (i.e. the RUN instruction in the Dockerfile)
|
||||
#
|
||||
# shellcheck disable=SC2068
|
||||
exec $@
|
||||
exec "$@"
|
||||
|
8
docker/housekeeping.sh
Executable file
8
docker/housekeeping.sh
Executable file
@ -0,0 +1,8 @@
|
||||
#!/bin/bash
|
||||
SECONDS=${HOUSEKEEPING_INTERVAL:=86400}
|
||||
echo "Interval set to ${SECONDS} seconds"
|
||||
while true; do
|
||||
date
|
||||
/opt/netbox/venv/bin/python /opt/netbox/netbox/manage.py housekeeping
|
||||
sleep "${SECONDS}s"
|
||||
done
|
@ -51,4 +51,6 @@ exec unitd \
|
||||
--pid /opt/unit/unit.pid \
|
||||
--log /dev/stdout \
|
||||
--state /opt/unit/state/ \
|
||||
--tmp /opt/unit/tmp/
|
||||
--tmp /opt/unit/tmp/ \
|
||||
--user unit \
|
||||
--group root
|
||||
|
@ -11,7 +11,7 @@
|
||||
"uri": "/static/*"
|
||||
},
|
||||
"action": {
|
||||
"share": "/opt/netbox/netbox"
|
||||
"share": "/opt/netbox/netbox${uri}"
|
||||
}
|
||||
},
|
||||
|
||||
|
4
env/netbox.env
vendored
4
env/netbox.env
vendored
@ -14,6 +14,8 @@ EMAIL_USERNAME=netbox
|
||||
# EMAIL_USE_SSL and EMAIL_USE_TLS are mutually exclusive, i.e. they can't both be `true`!
|
||||
EMAIL_USE_SSL=false
|
||||
EMAIL_USE_TLS=false
|
||||
GRAPHQL_ENABLED=true
|
||||
HOUSEKEEPING_INTERVAL=86400
|
||||
MAX_PAGE_SIZE=1000
|
||||
MEDIA_ROOT=/opt/netbox/netbox/media
|
||||
METRICS_ENABLED=false
|
||||
@ -22,10 +24,12 @@ NAPALM_TIMEOUT=10
|
||||
NAPALM_USERNAME=
|
||||
REDIS_CACHE_DATABASE=1
|
||||
REDIS_CACHE_HOST=redis-cache
|
||||
REDIS_CACHE_INSECURE_SKIP_TLS_VERIFY=false
|
||||
REDIS_CACHE_PASSWORD=t4Ph722qJ5QHeQ1qfu36
|
||||
REDIS_CACHE_SSL=false
|
||||
REDIS_DATABASE=0
|
||||
REDIS_HOST=redis
|
||||
REDIS_INSECURE_SKIP_TLS_VERIFY=false
|
||||
REDIS_PASSWORD=H733Kdjndks81
|
||||
REDIS_SSL=false
|
||||
RELEASE_CHECK_URL=https://api.github.com/repos/netbox-community/netbox/releases
|
||||
|
7
initializers/asns.yml
Normal file
7
initializers/asns.yml
Normal file
@ -0,0 +1,7 @@
|
||||
# - asn: 1
|
||||
# rir: RFC1918
|
||||
# tenant: tenant1
|
||||
# - asn: 2
|
||||
# rir: RFC4193 ULA
|
||||
# - asn: 3
|
||||
# rir: RFC3849
|
71
initializers/cables.yml
Normal file
71
initializers/cables.yml
Normal file
@ -0,0 +1,71 @@
|
||||
# # Required parameters for termination X ('a' or 'b'):
|
||||
# #
|
||||
# # ```
|
||||
# # termination_x_name -> name of interface
|
||||
# # termination_x_device -> name of the device interface belongs to
|
||||
# # termination_x_class -> required if different than 'Interface' which is the default
|
||||
# # ```
|
||||
# #
|
||||
# # Supported termination classes: Interface, ConsolePort, ConsoleServerPort, FrontPort, RearPort, PowerPort, PowerOutlet
|
||||
# #
|
||||
# #
|
||||
# # If a termination is a circuit then the required parameter is termination_x_circuit.
|
||||
# # Required parameters for a circuit termination:
|
||||
# #
|
||||
# # ```
|
||||
# # termination_x_circuit:
|
||||
# # term_side -> termination side of a circuit. Must be A or B
|
||||
# # cid -> circuit ID value
|
||||
# # site OR provider_network -> name of Site or ProviderNetwork respectively. If both provided, Site takes precedence
|
||||
# # ```
|
||||
# #
|
||||
# # If a termination is a power feed then the required parameter is termination_x_feed.
|
||||
# #
|
||||
# # ```
|
||||
# # termination_x_feed:
|
||||
# # name -> name of the PowerFeed object
|
||||
# # power_panel:
|
||||
# # name -> name of the PowerPanel the PowerFeed is attached to
|
||||
# # site -> name of the Site in which the PowerPanel is present
|
||||
# # ```
|
||||
# #
|
||||
# # Any other Cable parameters supported by Netbox are supported as the top level keys, e.g. 'type', 'status', etc.
|
||||
# #
|
||||
# # - termination_a_name: console
|
||||
# # termination_a_device: spine
|
||||
# # termination_a_class: ConsolePort
|
||||
# # termination_b_name: tty9
|
||||
# # termination_b_device: console-server
|
||||
# # termination_b_class: ConsoleServerPort
|
||||
# # type: cat6
|
||||
# #
|
||||
# - termination_a_name: to-server02
|
||||
# termination_a_device: server01
|
||||
# termination_b_name: to-server01
|
||||
# termination_b_device: server02
|
||||
# status: planned
|
||||
# type: mmf
|
||||
|
||||
# - termination_a_name: eth0
|
||||
# termination_a_device: server02
|
||||
# termination_b_circuit:
|
||||
# term_side: A
|
||||
# cid: Circuit_ID-1
|
||||
# site: AMS 1
|
||||
# type: cat6
|
||||
|
||||
# - termination_a_name: psu0
|
||||
# termination_a_device: server04
|
||||
# termination_a_class: PowerPort
|
||||
# termination_b_feed:
|
||||
# name: power feed 1
|
||||
# power_panel:
|
||||
# name: power panel AMS 1
|
||||
# site: AMS 1
|
||||
|
||||
# - termination_a_name: outlet1
|
||||
# termination_a_device: server04
|
||||
# termination_a_class: PowerOutlet
|
||||
# termination_b_name: psu1
|
||||
# termination_b_device: server04
|
||||
# termination_b_class: PowerPort
|
7
initializers/contact_groups.yml
Normal file
7
initializers/contact_groups.yml
Normal file
@ -0,0 +1,7 @@
|
||||
# - name: Network-Team
|
||||
# slug: network-team
|
||||
# description: This is a new contact group for the Network-Team
|
||||
# - name: New Contact Group
|
||||
# slug: new-contact-group
|
||||
# description: This is a new contact group sub under of Network-Team
|
||||
# parent: Network-Team
|
3
initializers/contact_roles.yml
Normal file
3
initializers/contact_roles.yml
Normal file
@ -0,0 +1,3 @@
|
||||
# - name: New Contact Role
|
||||
# slug: new-contact-role
|
||||
# description: This is a new contact role description
|
20
initializers/contacts.yml
Normal file
20
initializers/contacts.yml
Normal file
@ -0,0 +1,20 @@
|
||||
# - name: Lee Widget
|
||||
# title: CEO of Widget Corp
|
||||
# phone: 221-555-1212
|
||||
# email: widgetCEO@widgetcorp.com
|
||||
# address: 1200 Nowhere Blvd, Scranton NJ, 555111
|
||||
# comments: This is a very important contact
|
||||
# - name: Ali Gator
|
||||
# group: Network-Team
|
||||
# title: Consultant for Widget Corp
|
||||
# phone: 221-555-1213
|
||||
# email: Consultant@widgetcorp.com
|
||||
# address: 1200 Nowhere Blvd, Scranton NJ, 555111
|
||||
# comments: This is a very important contact
|
||||
# - name: Karlchen Maier
|
||||
# group: New Contact Group
|
||||
# title: COO of Widget Corp
|
||||
# phone: 221-555-1214
|
||||
# email: Karlchen@widgetcorp.com
|
||||
# address: 1200 Nowhere Blvd, Scranton NJ, 555111
|
||||
# comments: This is a very important contact
|
@ -10,12 +10,12 @@
|
||||
## Examples:
|
||||
|
||||
# - name: link_to_repo
|
||||
# text: 'Link to Netbox Docker'
|
||||
# url: 'https://github.com/netbox-community/netbox-docker'
|
||||
# link_text: 'Link to Netbox Docker'
|
||||
# link_url: 'https://github.com/netbox-community/netbox-docker'
|
||||
# new_window: False
|
||||
# content_type: device
|
||||
# - name: link_to_localhost
|
||||
# text: 'Link to localhost'
|
||||
# url: 'http://localhost'
|
||||
# link_text: 'Link to localhost'
|
||||
# link_url: 'http://localhost'
|
||||
# new_window: True
|
||||
# content_type: device
|
||||
|
@ -8,11 +8,28 @@
|
||||
##
|
||||
## Examples:
|
||||
|
||||
# - device: server01
|
||||
# name: ath0
|
||||
# type: 1000base-t
|
||||
# lag: ae0
|
||||
# bridge: br0
|
||||
# - device: server01
|
||||
# name: ath1
|
||||
# type: 1000base-t
|
||||
# parent: ath0
|
||||
# - device: server01
|
||||
# enabled: true
|
||||
# type: virtual
|
||||
# type: 1000base-x-sfp
|
||||
# name: to-server02
|
||||
# - device: server02
|
||||
# enabled: true
|
||||
# type: virtual
|
||||
# type: 1000base-x-sfp
|
||||
# name: to-server01
|
||||
# - device: server02
|
||||
# enabled: true
|
||||
# type: 1000base-t
|
||||
# name: eth0
|
||||
# - device: server02
|
||||
# enabled: true
|
||||
# type: virtual
|
||||
# name: loopback
|
||||
|
@ -21,3 +21,37 @@
|
||||
# slug: other
|
||||
# custom_field_data:
|
||||
# text_field: Description
|
||||
# interfaces:
|
||||
# - name: eth0
|
||||
# type: 1000base-t
|
||||
# mgmt_only: True
|
||||
# - name: eth1
|
||||
# type: 1000base-t
|
||||
# console_server_ports:
|
||||
# - name_template: ttyS[1-48]
|
||||
# type: rj-45
|
||||
# power_ports:
|
||||
# - name_template: psu[0,1]
|
||||
# type: iec-60320-c14
|
||||
# maximum_draw: 35
|
||||
# allocated_draw: 35
|
||||
# front_ports:
|
||||
# - name_template: front[1,2]
|
||||
# type: 8p8c
|
||||
# rear_port_template: rear[0,1]
|
||||
# rear_port_position_template: "[1,2]"
|
||||
# rear_ports:
|
||||
# - name_template: rear[0,1]
|
||||
# type: 8p8c
|
||||
# positions_template: "[3,2]"
|
||||
# device_bays:
|
||||
# - name: bay0 # both non-template and template field specified; non-template field takes precedence
|
||||
# name_template: bay[0-9]
|
||||
# label: test0
|
||||
# label_template: test[0-5,9,6-8]
|
||||
# description: Test description
|
||||
# power_outlets:
|
||||
# - name_template: outlet[0,1]
|
||||
# type: iec-60320-c5
|
||||
# power_port: psu0
|
||||
# feed_leg: B
|
||||
|
@ -42,3 +42,12 @@
|
||||
# position: 3
|
||||
# custom_field_data:
|
||||
# text_field: Description
|
||||
# - name: server04
|
||||
# device_role: server
|
||||
# device_type: Other
|
||||
# site: SING 1
|
||||
# location: cage 101
|
||||
# face: front
|
||||
# position: 3
|
||||
# custom_field_data:
|
||||
# text_field: Description
|
||||
|
@ -1,35 +1,9 @@
|
||||
## To list all permissions, run:
|
||||
##
|
||||
## docker-compose run --rm --entrypoint /bin/bash netbox
|
||||
## $ ./manage.py migrate
|
||||
## $ ./manage.py shell
|
||||
## > from django.contrib.auth.models import Permission
|
||||
## > print('\n'.join([p.codename for p in Permission.objects.all()]))
|
||||
##
|
||||
## Permission lists support wildcards. See the examples below.
|
||||
##
|
||||
## Examples:
|
||||
|
||||
# applications:
|
||||
# users:
|
||||
# - technical_user
|
||||
# - technical_user
|
||||
# readers:
|
||||
# users:
|
||||
# - reader
|
||||
# - reader
|
||||
# writers:
|
||||
# users:
|
||||
# - writer
|
||||
# permissions:
|
||||
# - delete_device
|
||||
# - delete_virtualmachine
|
||||
# - add_*
|
||||
# - change_*
|
||||
# vm_managers:
|
||||
# permissions:
|
||||
# - '*_virtualmachine'
|
||||
# device_managers:
|
||||
# permissions:
|
||||
# - '*device*'
|
||||
# creators:
|
||||
# permissions:
|
||||
# - add_*
|
||||
# - writer
|
||||
|
48
initializers/object_permissions.yml
Normal file
48
initializers/object_permissions.yml
Normal file
@ -0,0 +1,48 @@
|
||||
# all.ro:
|
||||
# actions:
|
||||
# - view
|
||||
# description: 'Read Only for All Objects'
|
||||
# enabled: true
|
||||
# groups:
|
||||
# - applications
|
||||
# - readers
|
||||
# object_types: all
|
||||
# users:
|
||||
# - jdoe
|
||||
# all.rw:
|
||||
# actions:
|
||||
# - add
|
||||
# - change
|
||||
# - delete
|
||||
# - view
|
||||
# description: 'Read/Write for All Objects'
|
||||
# enabled: true
|
||||
# groups:
|
||||
# - writers
|
||||
# object_types: all
|
||||
# network_team.rw:
|
||||
# actions:
|
||||
# - add
|
||||
# - change
|
||||
# - delete
|
||||
# - view
|
||||
# description: "Network Team Permissions"
|
||||
# enabled: true
|
||||
# object_types:
|
||||
# circuits:
|
||||
# - circuit
|
||||
# - circuittermination
|
||||
# - circuittype
|
||||
# - provider
|
||||
# dcim: all
|
||||
# ipam:
|
||||
# - aggregate
|
||||
# - ipaddress
|
||||
# - prefix
|
||||
# - rir
|
||||
# - role
|
||||
# - routetarget
|
||||
# - service
|
||||
# - vlan
|
||||
# - vlangroup
|
||||
# - vrf
|
@ -2,4 +2,4 @@
|
||||
# site: AMS 1
|
||||
# - name: power panel SING 1
|
||||
# site: SING 1
|
||||
# rack_group: cage 101
|
||||
# location: cage 101
|
||||
|
@ -32,7 +32,7 @@
|
||||
# text_field: Description
|
||||
# - site: SING 1
|
||||
# name: rack-03
|
||||
# group: cage 101
|
||||
# location: cage 101
|
||||
# role: Role 3
|
||||
# type: 4-post-cabinet
|
||||
# width: 19
|
||||
|
@ -1,4 +0,0 @@
|
||||
# - name: Super Secret Passwords
|
||||
# slug: super-secret
|
||||
# - name: SNMP Communities
|
||||
# slug: snmp
|
@ -3,7 +3,6 @@
|
||||
# region: Downtown
|
||||
# status: active
|
||||
# facility: Amsterdam 1
|
||||
# asn: 12345
|
||||
# custom_field_data:
|
||||
# text_field: Description for AMS1
|
||||
# - name: AMS 2
|
||||
@ -11,7 +10,6 @@
|
||||
# region: Downtown
|
||||
# status: active
|
||||
# facility: Amsterdam 2
|
||||
# asn: 54321
|
||||
# custom_field_data:
|
||||
# text_field: Description for AMS2
|
||||
# - name: AMS 3
|
||||
@ -19,7 +17,7 @@
|
||||
# region: Suburbs
|
||||
# status: active
|
||||
# facility: Amsterdam 3
|
||||
# asn: 67890
|
||||
# tenant: tenant1
|
||||
# custom_field_data:
|
||||
# text_field: Description for AMS3
|
||||
# - name: SING 1
|
||||
@ -27,6 +25,6 @@
|
||||
# region: Singapore
|
||||
# status: active
|
||||
# facility: Singapore 1
|
||||
# asn: 09876
|
||||
# tenant: tenant2
|
||||
# custom_field_data:
|
||||
# text_field: Description for SING1
|
||||
|
@ -1,23 +1,15 @@
|
||||
## To list all permissions, run:
|
||||
##
|
||||
## docker-compose run --rm --entrypoint /bin/bash netbox
|
||||
## $ ./manage.py migrate
|
||||
## $ ./manage.py shell
|
||||
## > from django.contrib.auth.models import Permission
|
||||
## > print('\n'.join([p.codename for p in Permission.objects.all()]))
|
||||
##
|
||||
## Permission lists support wildcards. See the examples below.
|
||||
##
|
||||
## Examples:
|
||||
|
||||
# technical_user:
|
||||
# api_token: 0123456789technicaluser789abcdef01234567 # must be looooong!
|
||||
# reader:
|
||||
# password: reader
|
||||
# writer:
|
||||
# password: writer
|
||||
# permissions:
|
||||
# - delete_device
|
||||
# - delete_virtualmachine
|
||||
# - add_*
|
||||
# - change_*
|
||||
# api_token: "" # a token is generated automatically unless the value is explicity set to empty
|
||||
# jdoe:
|
||||
# first_name: John
|
||||
# last_name: Doe
|
||||
# api_token: 0123456789jdoe789abcdef01234567jdoe
|
||||
# is_active: True
|
||||
# is_superuser: False
|
||||
# is_staff: False
|
||||
# email: john.doe@example.com
|
||||
|
@ -1,6 +1,24 @@
|
||||
# - name: VLAN group 1
|
||||
# site: AMS 1
|
||||
# scope_type: dcim.region
|
||||
# scope: Amsterdam
|
||||
# slug: vlan-group-1
|
||||
# - name: VLAN group 2
|
||||
# site: AMS 1
|
||||
# scope_type: dcim.site
|
||||
# scope: AMS 1
|
||||
# slug: vlan-group-2
|
||||
# - name: VLAN group 3
|
||||
# scope_type: dcim.location
|
||||
# scope: cage 101
|
||||
# slug: vlan-group-3
|
||||
# - name: VLAN group 4
|
||||
# scope_type: dcim.rack
|
||||
# scope: rack-01
|
||||
# slug: vlan-group-4
|
||||
# - name: VLAN group 5
|
||||
# scope_type: virtualization.cluster
|
||||
# scope: cluster1
|
||||
# slug: vlan-group-5
|
||||
# - name: VLAN group 6
|
||||
# scope_type: virtualization.clustergroup
|
||||
# scope: Group 1
|
||||
# slug: vlan-group-6
|
||||
|
188
release.sh
Executable file
188
release.sh
Executable file
@ -0,0 +1,188 @@
|
||||
#!/bin/bash
|
||||
|
||||
DEFAULT_REPO=netbox-community/netbox-docker
|
||||
REPO="${REPO-${DEFAULT_REPO}}"
|
||||
|
||||
echomoji() {
|
||||
EMOJI=${1}
|
||||
TEXT=${2}
|
||||
shift 2
|
||||
if [ -z "$DISABLE_EMOJI" ]; then
|
||||
echo "${EMOJI}" "${@}"
|
||||
else
|
||||
echo "${TEXT}" "${@}"
|
||||
fi
|
||||
}
|
||||
|
||||
echo_nok() {
|
||||
echomoji "❌" "!" "${@}"
|
||||
}
|
||||
echo_ok() {
|
||||
echomoji "✅" "-" "${@}"
|
||||
}
|
||||
echo_hint() {
|
||||
echomoji "👉" ">" "${@}"
|
||||
}
|
||||
|
||||
# check errors shall exit with code 1
|
||||
|
||||
check_clean_repo() {
|
||||
changes=$(git status --porcelain 2>/dev/null)
|
||||
if [ ${?} ] && [ -n "$changes" ]; then
|
||||
echo_nok "There are git changes pending:"
|
||||
echo "$changes"
|
||||
echo_hint "Please clean the repository before continueing: git stash --include-untracked"
|
||||
exit 1
|
||||
fi
|
||||
echo_ok "Repository has no pending changes."
|
||||
}
|
||||
|
||||
check_branch() {
|
||||
expected_branch="${1}"
|
||||
actual_branch=$(git rev-parse --abbrev-ref HEAD 2>/dev/null)
|
||||
if [ ${?} ] && [ "${actual_branch}" != "${expected_branch}" ]; then
|
||||
echo_nok "Current branch should be '${expected_branch}', but is '${actual_branch}'."
|
||||
echo_hint "Please change to the '${expected_branch}' branch: git checkout ${expected_branch}"
|
||||
exit 1
|
||||
fi
|
||||
echo_ok "The current branch is '${actual_branch}'."
|
||||
}
|
||||
|
||||
check_upstream() {
|
||||
expected_upstream_branch="origin/${1}"
|
||||
actual_upstream_branch=$(git rev-parse --abbrev-ref '@{upstream}' 2>/dev/null)
|
||||
if [ ${?} ] && [ "${actual_upstream_branch}" != "${expected_upstream_branch}" ]; then
|
||||
echo_nok "Current upstream branch should be '${expected_upstream_branch}', but is '${actual_upstream_branch}'."
|
||||
echo_hint "Please set '${expected_upstream_branch}' as the upstream branch: git branch --set-upstream-to=${expected_upstream_branch}"
|
||||
exit 1
|
||||
fi
|
||||
echo_ok "The current upstream branch is '${actual_upstream_branch}'."
|
||||
}
|
||||
|
||||
check_origin() {
|
||||
expected_origin="git@github.com:${REPO}.git"
|
||||
actual_origin=$(git remote get-url origin 2>/dev/null)
|
||||
if [ ${?} ] && [ "${actual_origin}" != "${expected_origin}" ]; then
|
||||
echo_nok "The url of origin is '${actual_origin}', but '${expected_origin}' is expected."
|
||||
echo_hint "Please set '${expected_origin}' as the url for origin: git origin set-url '${expected_origin}'"
|
||||
exit 1
|
||||
fi
|
||||
echo_ok "The current origin url is '${actual_origin}'."
|
||||
}
|
||||
|
||||
check_latest() {
|
||||
git fetch --tags origin
|
||||
|
||||
local_head_commit=$(git rev-parse HEAD 2>/dev/null)
|
||||
remote_head_commit=$(git rev-parse FETCH_HEAD 2>/dev/null)
|
||||
if [ "${local_head_commit}" != "${remote_head_commit}" ]; then
|
||||
echo_nok "HEAD is at '${local_head_commit}', but FETCH_HEAD is at '${remote_head_commit}'."
|
||||
echo_hint "Please ensure that you have pushed and pulled all the latest chanegs: git pull --prune --rebase origin; git push origin"
|
||||
exit 1
|
||||
fi
|
||||
echo_ok "HEAD and FETCH_HEAD both point to '${local_head_commit}'."
|
||||
}
|
||||
|
||||
check_tag() {
|
||||
local tag
|
||||
|
||||
tag=$(<VERSION)
|
||||
if git rev-parse "${tag}" 2>/dev/null >/dev/null; then
|
||||
echo_nok "The tag '${tag}' already points to '$(git rev-parse "${tag}" 2>/dev/null)'."
|
||||
echo_hint "Please ensure that the 'VERSION' file has been updated before trying to release: echo X.Y.Z > VERSION"
|
||||
exit 1
|
||||
fi
|
||||
echo_ok "The tag '${tag}' does not exist yet."
|
||||
}
|
||||
|
||||
check_develop() {
|
||||
echomoji 📋 "?" "Checking 'develop' branch"
|
||||
|
||||
check_branch develop
|
||||
check_upstream develop
|
||||
check_clean_repo
|
||||
check_latest
|
||||
}
|
||||
|
||||
check_release() {
|
||||
echomoji 📋 "?" "Checking 'release' branch"
|
||||
|
||||
check_upstream release
|
||||
check_clean_repo
|
||||
check_latest
|
||||
}
|
||||
|
||||
# git errors shall exit with code 2
|
||||
|
||||
git_switch() {
|
||||
echomoji 🔀 "≈" "Switching to '${1}' branch…"
|
||||
if ! git checkout "${1}" >/dev/null; then
|
||||
echo_nok "It was not possible to switch to the branch '${1}'."
|
||||
exit 2
|
||||
fi
|
||||
echo_ok "The branch is now '${1}'."
|
||||
}
|
||||
|
||||
git_tag() {
|
||||
echomoji 🏷 "X" "Tagging version '${1}'…"
|
||||
if ! git tag "${1}"; then
|
||||
echo_nok "The tag '${1}' was not created because of an error."
|
||||
exit 2
|
||||
fi
|
||||
echo_ok "The tag '$(<VERSION)' was created."
|
||||
}
|
||||
|
||||
git_push() {
|
||||
echomoji ⏩ "»" "Pushing the tag '${2}' to '${1}'…"
|
||||
if ! git push "${1}" "${2}"; then
|
||||
echo_nok "The tag '${2}' could not be pushed to '${1}'."
|
||||
exit 2
|
||||
fi
|
||||
echo_ok "The tag '${2}' was pushed."
|
||||
}
|
||||
|
||||
git_merge() {
|
||||
echomoji ⏩ "»" "Merging '${1}'…"
|
||||
if ! git merge --no-ff "${1}"; then
|
||||
echo_nok "The branch '${1}' could not be merged."
|
||||
exit 2
|
||||
fi
|
||||
echo_ok "The branch '${2}' was merged."
|
||||
}
|
||||
|
||||
git_merge() {
|
||||
echomoji ⏩ "»" "Rebasing onto '${1}'…"
|
||||
if ! git rebase "${1}"; then
|
||||
echo_nok "Could not rebase onto '${1}'."
|
||||
exit 2
|
||||
fi
|
||||
echo_ok "Rebased onto '${2}'."
|
||||
}
|
||||
|
||||
###
|
||||
# MAIN
|
||||
###
|
||||
|
||||
echomoji 📋 "▶︎" "Checking pre-requisites for releasing '$(<VERSION)'"
|
||||
|
||||
check_origin
|
||||
|
||||
check_develop
|
||||
check_tag
|
||||
|
||||
git_switch release
|
||||
check_release
|
||||
|
||||
echomoji 📋 "▶︎" "Releasing '$(<VERSION)'"
|
||||
|
||||
git_merge develop
|
||||
check_tag
|
||||
git_tag "$(<VERSION)"
|
||||
|
||||
git_push "origin" release
|
||||
git_push "origin" "$(<VERSION)"
|
||||
|
||||
git_switch develop
|
||||
git_rebase release
|
||||
|
||||
echomoji ✅ "◼︎" "The release of '$(<VERSION)' is complete."
|
@ -1,6 +1,7 @@
|
||||
{
|
||||
"extends": [
|
||||
"config:base"
|
||||
"config:base",
|
||||
":disableDependencyDashboard"
|
||||
],
|
||||
"enabled": true,
|
||||
"labels": ["maintenance"],
|
||||
|
@ -1,4 +1,6 @@
|
||||
napalm==3.2.0
|
||||
ruamel.yaml==0.16.12
|
||||
django-auth-ldap==2.2.0
|
||||
django-storages[azure,boto3,dropbox,google,libcloud,sftp]==1.11.1
|
||||
django-auth-ldap==4.1.0
|
||||
django-storages[azure,boto3,dropbox,google,libcloud,sftp]==1.12.3
|
||||
napalm==4.0.0
|
||||
psycopg2==2.9.3
|
||||
social-auth-core[openidconnect]==4.3.0
|
||||
ruamel.yaml==0.17.21
|
||||
|
@ -1,7 +1,7 @@
|
||||
import sys
|
||||
|
||||
from django.contrib.auth.models import User
|
||||
from startup_script_utils import load_yaml, set_permissions
|
||||
from startup_script_utils import load_yaml
|
||||
from users.models import Token
|
||||
|
||||
users = load_yaml("/opt/netbox/initializers/users.yml")
|
||||
@ -9,16 +9,17 @@ if users is None:
|
||||
sys.exit()
|
||||
|
||||
for username, user_details in users.items():
|
||||
if not User.objects.filter(username=username):
|
||||
user = User.objects.create_user(
|
||||
username=username,
|
||||
password=user_details.get("password", 0) or User.objects.make_random_password(),
|
||||
)
|
||||
|
||||
api_token = user_details.pop("api_token", Token.generate_key())
|
||||
password = user_details.pop("password", User.objects.make_random_password())
|
||||
|
||||
user, created = User.objects.get_or_create(username=username, defaults=user_details)
|
||||
|
||||
if created:
|
||||
user.set_password(password)
|
||||
user.save()
|
||||
|
||||
if api_token:
|
||||
Token.objects.get_or_create(user=user, key=api_token)
|
||||
|
||||
print("👤 Created user", username)
|
||||
|
||||
if user_details.get("api_token", 0):
|
||||
Token.objects.create(user=user, key=user_details["api_token"])
|
||||
|
||||
yaml_permissions = user_details.get("permissions", [])
|
||||
set_permissions(user.user_permissions, yaml_permissions)
|
||||
|
@ -1,23 +1,23 @@
|
||||
import sys
|
||||
|
||||
from django.contrib.auth.models import Group, User
|
||||
from startup_script_utils import load_yaml, set_permissions
|
||||
from startup_script_utils import load_yaml
|
||||
from users.models import AdminGroup, AdminUser
|
||||
|
||||
groups = load_yaml("/opt/netbox/initializers/groups.yml")
|
||||
if groups is None:
|
||||
sys.exit()
|
||||
|
||||
for groupname, group_details in groups.items():
|
||||
group, created = Group.objects.get_or_create(name=groupname)
|
||||
group, created = AdminGroup.objects.get_or_create(name=groupname)
|
||||
|
||||
if created:
|
||||
print("👥 Created group", groupname)
|
||||
|
||||
for username in group_details.get("users", []):
|
||||
user = User.objects.get(username=username)
|
||||
user = AdminUser.objects.get(username=username)
|
||||
|
||||
if user:
|
||||
user.groups.add(group)
|
||||
group.user_set.add(user)
|
||||
print(" 👤 Assigned user %s to group %s" % (username, group.name))
|
||||
|
||||
yaml_permissions = group_details.get("permissions", [])
|
||||
set_permissions(group.permissions, yaml_permissions)
|
||||
group.save()
|
||||
|
68
startup_scripts/020_object_permissions.py
Normal file
68
startup_scripts/020_object_permissions.py
Normal file
@ -0,0 +1,68 @@
|
||||
import sys
|
||||
|
||||
from django.contrib.contenttypes.models import ContentType
|
||||
from startup_script_utils import load_yaml
|
||||
from users.models import AdminGroup, AdminUser, ObjectPermission
|
||||
|
||||
object_permissions = load_yaml("/opt/netbox/initializers/object_permissions.yml")
|
||||
|
||||
if object_permissions is None:
|
||||
sys.exit()
|
||||
|
||||
|
||||
for permission_name, permission_details in object_permissions.items():
|
||||
|
||||
object_permission, created = ObjectPermission.objects.get_or_create(
|
||||
name=permission_name,
|
||||
defaults={
|
||||
"description": permission_details["description"],
|
||||
"enabled": permission_details["enabled"],
|
||||
"actions": permission_details["actions"],
|
||||
},
|
||||
)
|
||||
|
||||
if permission_details.get("object_types", 0):
|
||||
object_types = permission_details["object_types"]
|
||||
|
||||
if object_types == "all":
|
||||
object_permission.object_types.set(ContentType.objects.all())
|
||||
|
||||
else:
|
||||
for app_label, models in object_types.items():
|
||||
if models == "all":
|
||||
app_models = ContentType.objects.filter(app_label=app_label)
|
||||
|
||||
for app_model in app_models:
|
||||
object_permission.object_types.add(app_model.id)
|
||||
else:
|
||||
# There is
|
||||
for model in models:
|
||||
object_permission.object_types.add(
|
||||
ContentType.objects.get(app_label=app_label, model=model)
|
||||
)
|
||||
|
||||
print("🔓 Created object permission", object_permission.name)
|
||||
|
||||
if permission_details.get("groups", 0):
|
||||
for groupname in permission_details["groups"]:
|
||||
group = AdminGroup.objects.filter(name=groupname).first()
|
||||
|
||||
if group:
|
||||
object_permission.groups.add(group)
|
||||
print(
|
||||
" 👥 Assigned group %s object permission of %s"
|
||||
% (groupname, object_permission.name)
|
||||
)
|
||||
|
||||
if permission_details.get("users", 0):
|
||||
for username in permission_details["users"]:
|
||||
user = AdminUser.objects.filter(username=username).first()
|
||||
|
||||
if user:
|
||||
object_permission.users.add(user)
|
||||
print(
|
||||
" 👤 Assigned user %s object permission of %s"
|
||||
% (username, object_permission.name)
|
||||
)
|
||||
|
||||
object_permission.save()
|
@ -42,6 +42,9 @@ for cf_name, cf_details in customfields.items():
|
||||
if cf_details.get("type", False):
|
||||
custom_field.type = cf_details["type"]
|
||||
|
||||
if cf_details.get("filter_logic", False):
|
||||
custom_field.filter_logic = cf_details["filter_logic"]
|
||||
|
||||
if cf_details.get("weight", -1) >= 0:
|
||||
custom_field.weight = cf_details["weight"]
|
||||
|
@ -2,7 +2,7 @@ import sys
|
||||
|
||||
from django.contrib.contenttypes.models import ContentType
|
||||
from extras.models import CustomLink
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
|
||||
custom_links = load_yaml("/opt/netbox/initializers/custom_links.yml")
|
||||
|
||||
@ -23,11 +23,13 @@ for link in custom_links:
|
||||
if link["content_type_id"] is None:
|
||||
print(
|
||||
"⚠️ Unable to create Custom Link '{0}': The content_type '{1}' is unknown".format(
|
||||
link.name, content_type
|
||||
link.get("name"), content_type
|
||||
)
|
||||
)
|
||||
continue
|
||||
|
||||
custom_link, created = CustomLink.objects.get_or_create(**link)
|
||||
matching_params, defaults = split_params(link)
|
||||
custom_link, created = CustomLink.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🔗 Created Custom Link '{0}'".format(custom_link.name))
|
@ -1,7 +1,7 @@
|
||||
import sys
|
||||
|
||||
from extras.models import Tag
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
from utilities.choices import ColorChoices
|
||||
|
||||
tags = load_yaml("/opt/netbox/initializers/tags.yml")
|
||||
@ -17,7 +17,8 @@ for params in tags:
|
||||
if color in color_tpl:
|
||||
params["color"] = color_tpl[0]
|
||||
|
||||
tag, created = Tag.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
tag, created = Tag.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🎨 Created Tag", tag.name)
|
@ -1,37 +0,0 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import DeviceType, Manufacturer, Region
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from tenancy.models import Tenant
|
||||
|
||||
device_types = load_yaml("/opt/netbox/initializers/device_types.yml")
|
||||
|
||||
if device_types is None:
|
||||
sys.exit()
|
||||
|
||||
required_assocs = {"manufacturer": (Manufacturer, "name")}
|
||||
|
||||
optional_assocs = {"region": (Region, "name"), "tenant": (Tenant, "name")}
|
||||
|
||||
for params in device_types:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
|
||||
for assoc, details in required_assocs.items():
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
for assoc, details in optional_assocs.items():
|
||||
if assoc in params:
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
device_type, created = DeviceType.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(device_type, custom_field_data)
|
||||
|
||||
print("🔡 Created device type", device_type.manufacturer, device_type.model)
|
@ -2,7 +2,7 @@ import sys
|
||||
|
||||
from django.contrib.contenttypes.models import ContentType
|
||||
from extras.models import Webhook
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
|
||||
webhooks = load_yaml("/opt/netbox/initializers/webhooks.yml")
|
||||
|
||||
@ -26,7 +26,9 @@ for hook in webhooks:
|
||||
except ContentType.DoesNotExist:
|
||||
continue
|
||||
|
||||
webhook, created = Webhook.objects.get_or_create(**hook)
|
||||
matching_params, defaults = split_params(hook)
|
||||
webhook, created = Webhook.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
webhook.content_types.set(obj_type_ids)
|
||||
webhook.save()
|
@ -1,6 +1,6 @@
|
||||
import sys
|
||||
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
from tenancy.models import TenantGroup
|
||||
|
||||
tenant_groups = load_yaml("/opt/netbox/initializers/tenant_groups.yml")
|
||||
@ -9,7 +9,8 @@ if tenant_groups is None:
|
||||
sys.exit()
|
||||
|
||||
for params in tenant_groups:
|
||||
tenant_group, created = TenantGroup.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
tenant_group, created = TenantGroup.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🔳 Created Tenant Group", tenant_group.name)
|
@ -1,23 +0,0 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import RackGroup, Site
|
||||
from startup_script_utils import load_yaml
|
||||
|
||||
rack_groups = load_yaml("/opt/netbox/initializers/rack_groups.yml")
|
||||
|
||||
if rack_groups is None:
|
||||
sys.exit()
|
||||
|
||||
required_assocs = {"site": (Site, "name")}
|
||||
|
||||
for params in rack_groups:
|
||||
|
||||
for assoc, details in required_assocs.items():
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
rack_group, created = RackGroup.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
||||
print("🎨 Created rack group", rack_group.name)
|
@ -1,6 +1,11 @@
|
||||
import sys
|
||||
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant, TenantGroup
|
||||
|
||||
tenants = load_yaml("/opt/netbox/initializers/tenants.yml")
|
||||
@ -20,9 +25,10 @@ for params in tenants:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
tenant, created = Tenant.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
tenant, created = Tenant.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(tenant, custom_field_data)
|
||||
|
||||
print("👩💻 Created Tenant", tenant.name)
|
||||
|
||||
set_custom_fields_values(tenant, custom_field_data)
|
@ -1,7 +1,7 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Region
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
|
||||
regions = load_yaml("/opt/netbox/initializers/regions.yml")
|
||||
|
||||
@ -19,7 +19,8 @@ for params in regions:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
region, created = Region.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
region, created = Region.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🌐 Created region", region.name)
|
@ -1,7 +1,12 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Region, Site
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
|
||||
sites = load_yaml("/opt/netbox/initializers/sites.yml")
|
||||
@ -21,9 +26,10 @@ for params in sites:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
site, created = Site.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
site, created = Site.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(site, custom_field_data)
|
||||
|
||||
print("📍 Created site", site.name)
|
||||
|
||||
set_custom_fields_values(site, custom_field_data)
|
25
startup_scripts/120_locations.py
Normal file
25
startup_scripts/120_locations.py
Normal file
@ -0,0 +1,25 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Location, Site
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
|
||||
rack_groups = load_yaml("/opt/netbox/initializers/locations.yml")
|
||||
|
||||
if rack_groups is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["name", "slug", "site"]
|
||||
required_assocs = {"site": (Site, "name")}
|
||||
|
||||
for params in rack_groups:
|
||||
|
||||
for assoc, details in required_assocs.items():
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
location, created = Location.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🎨 Created location", location.name)
|
@ -1,7 +1,7 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import RackRole
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
from utilities.choices import ColorChoices
|
||||
|
||||
rack_roles = load_yaml("/opt/netbox/initializers/rack_roles.yml")
|
||||
@ -17,7 +17,8 @@ for params in rack_roles:
|
||||
if color in color_tpl:
|
||||
params["color"] = color_tpl[0]
|
||||
|
||||
rack_role, created = RackRole.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
rack_role, created = RackRole.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🎨 Created rack role", rack_role.name)
|
@ -1,42 +0,0 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Site
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from tenancy.models import Tenant
|
||||
from virtualization.models import Cluster, ClusterGroup, ClusterType
|
||||
|
||||
clusters = load_yaml("/opt/netbox/initializers/clusters.yml")
|
||||
|
||||
if clusters is None:
|
||||
sys.exit()
|
||||
|
||||
required_assocs = {"type": (ClusterType, "name")}
|
||||
|
||||
optional_assocs = {
|
||||
"site": (Site, "name"),
|
||||
"group": (ClusterGroup, "name"),
|
||||
"tenant": (Tenant, "name"),
|
||||
}
|
||||
|
||||
for params in clusters:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
|
||||
for assoc, details in required_assocs.items():
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
for assoc, details in optional_assocs.items():
|
||||
if assoc in params:
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
cluster, created = Cluster.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(cluster, custom_field_data)
|
||||
|
||||
print("🗄️ Created cluster", cluster.name)
|
@ -1,51 +0,0 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Device, DeviceRole, DeviceType, Platform, Rack, Site
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from tenancy.models import Tenant
|
||||
from virtualization.models import Cluster
|
||||
|
||||
devices = load_yaml("/opt/netbox/initializers/devices.yml")
|
||||
|
||||
if devices is None:
|
||||
sys.exit()
|
||||
|
||||
required_assocs = {
|
||||
"device_role": (DeviceRole, "name"),
|
||||
"device_type": (DeviceType, "model"),
|
||||
"site": (Site, "name"),
|
||||
}
|
||||
|
||||
optional_assocs = {
|
||||
"tenant": (Tenant, "name"),
|
||||
"platform": (Platform, "name"),
|
||||
"rack": (Rack, "name"),
|
||||
"cluster": (Cluster, "name"),
|
||||
}
|
||||
|
||||
for params in devices:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
|
||||
# primary ips are handled later in `270_primary_ips.py`
|
||||
params.pop("primary_ip4", None)
|
||||
params.pop("primary_ip6", None)
|
||||
|
||||
for assoc, details in required_assocs.items():
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
for assoc, details in optional_assocs.items():
|
||||
if assoc in params:
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
device, created = Device.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(device, custom_field_data)
|
||||
|
||||
print("🖥️ Created device", device.name)
|
@ -1,7 +1,12 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Rack, RackGroup, RackRole, Site
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from dcim.models import Location, Rack, RackRole, Site
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
|
||||
racks = load_yaml("/opt/netbox/initializers/racks.yml")
|
||||
@ -9,12 +14,12 @@ racks = load_yaml("/opt/netbox/initializers/racks.yml")
|
||||
if racks is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["name", "site"]
|
||||
required_assocs = {"site": (Site, "name")}
|
||||
|
||||
optional_assocs = {
|
||||
"role": (RackRole, "name"),
|
||||
"tenant": (Tenant, "name"),
|
||||
"group": (RackGroup, "name"),
|
||||
"location": (Location, "name"),
|
||||
}
|
||||
|
||||
for params in racks:
|
||||
@ -33,9 +38,10 @@ for params in racks:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
rack, created = Rack.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
rack, created = Rack.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(rack, custom_field_data)
|
||||
|
||||
print("🔳 Created rack", rack.site, rack.name)
|
||||
|
||||
set_custom_fields_values(rack, custom_field_data)
|
@ -1,16 +1,21 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import PowerPanel, RackGroup, Site
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from dcim.models import Location, PowerPanel, Site
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
|
||||
power_panels = load_yaml("/opt/netbox/initializers/power_panels.yml")
|
||||
|
||||
if power_panels is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["name", "site"]
|
||||
required_assocs = {"site": (Site, "name")}
|
||||
|
||||
optional_assocs = {"rack_group": (RackGroup, "name")}
|
||||
optional_assocs = {"location": (Location, "name")}
|
||||
|
||||
for params in power_panels:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
@ -28,9 +33,10 @@ for params in power_panels:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
power_panel, created = PowerPanel.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
power_panel, created = PowerPanel.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(power_panel, custom_field_data)
|
||||
|
||||
print("⚡ Created Power Panel", power_panel.site, power_panel.name)
|
||||
|
||||
set_custom_fields_values(power_panel, custom_field_data)
|
@ -1,15 +1,20 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import PowerFeed, PowerPanel, Rack
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
|
||||
power_feeds = load_yaml("/opt/netbox/initializers/power_feeds.yml")
|
||||
|
||||
if power_feeds is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["name", "power_panel"]
|
||||
required_assocs = {"power_panel": (PowerPanel, "name")}
|
||||
|
||||
optional_assocs = {"rack": (Rack, "name")}
|
||||
|
||||
for params in power_feeds:
|
||||
@ -28,9 +33,10 @@ for params in power_feeds:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
power_feed, created = PowerFeed.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
power_feed, created = PowerFeed.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(power_feed, custom_field_data)
|
||||
|
||||
print("⚡ Created Power Feed", power_feed.name)
|
||||
|
||||
set_custom_fields_values(power_feed, custom_field_data)
|
@ -1,15 +0,0 @@
|
||||
import sys
|
||||
|
||||
from startup_script_utils import load_yaml
|
||||
from virtualization.models import ClusterGroup
|
||||
|
||||
cluster_groups = load_yaml("/opt/netbox/initializers/cluster_groups.yml")
|
||||
|
||||
if cluster_groups is None:
|
||||
sys.exit()
|
||||
|
||||
for params in cluster_groups:
|
||||
cluster_group, created = ClusterGroup.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
||||
print("🗄️ Created Cluster Group", cluster_group.name)
|
@ -1,7 +1,7 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Manufacturer
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
|
||||
manufacturers = load_yaml("/opt/netbox/initializers/manufacturers.yml")
|
||||
|
||||
@ -9,7 +9,8 @@ if manufacturers is None:
|
||||
sys.exit()
|
||||
|
||||
for params in manufacturers:
|
||||
manufacturer, created = Manufacturer.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
manufacturer, created = Manufacturer.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🏭 Created Manufacturer", manufacturer.name)
|
@ -1,7 +1,7 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import DeviceRole
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
from utilities.choices import ColorChoices
|
||||
|
||||
device_roles = load_yaml("/opt/netbox/initializers/device_roles.yml")
|
||||
@ -18,7 +18,8 @@ for params in device_roles:
|
||||
if color in color_tpl:
|
||||
params["color"] = color_tpl[0]
|
||||
|
||||
device_role, created = DeviceRole.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
device_role, created = DeviceRole.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🎨 Created device role", device_role.name)
|
139
startup_scripts/190_device_types.py
Normal file
139
startup_scripts/190_device_types.py
Normal file
@ -0,0 +1,139 @@
|
||||
import sys
|
||||
from typing import List
|
||||
|
||||
from dcim.models import DeviceType, Manufacturer, Region
|
||||
from dcim.models.device_component_templates import (
|
||||
ConsolePortTemplate,
|
||||
ConsoleServerPortTemplate,
|
||||
DeviceBayTemplate,
|
||||
FrontPortTemplate,
|
||||
InterfaceTemplate,
|
||||
PowerOutletTemplate,
|
||||
PowerPortTemplate,
|
||||
RearPortTemplate,
|
||||
)
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
from utilities.forms.utils import expand_alphanumeric_pattern
|
||||
|
||||
|
||||
def expand_templates(params: List[dict], device_type: DeviceType) -> List[dict]:
|
||||
templateable_fields = ["name", "label", "positions", "rear_port", "rear_port_position"]
|
||||
|
||||
expanded = []
|
||||
for param in params:
|
||||
param["device_type"] = device_type
|
||||
expanded_fields = {}
|
||||
has_plain_fields = False
|
||||
|
||||
for field in templateable_fields:
|
||||
template_value = param.pop(f"{field}_template", None)
|
||||
|
||||
if field in param:
|
||||
has_plain_fields = True
|
||||
elif template_value:
|
||||
expanded_fields[field] = list(expand_alphanumeric_pattern(template_value))
|
||||
|
||||
if expanded_fields and has_plain_fields:
|
||||
raise ValueError(f"Mix of plain and template keys provided for {templateable_fields}")
|
||||
elif not expanded_fields:
|
||||
expanded.append(param)
|
||||
continue
|
||||
|
||||
elements = list(expanded_fields.values())
|
||||
master_len = len(elements[0])
|
||||
if not all([len(elem) == master_len for elem in elements]):
|
||||
raise ValueError(
|
||||
f"Number of elements in template fields "
|
||||
f"{list(expanded_fields.keys())} must be equal"
|
||||
)
|
||||
|
||||
for idx in range(master_len):
|
||||
tmp = param.copy()
|
||||
for field, value in expanded_fields.items():
|
||||
if field in nested_assocs:
|
||||
model, match_key = nested_assocs[field]
|
||||
query = {match_key: value[idx], "device_type": device_type}
|
||||
tmp[field] = model.objects.get(**query)
|
||||
else:
|
||||
tmp[field] = value[idx]
|
||||
expanded.append(tmp)
|
||||
return expanded
|
||||
|
||||
|
||||
device_types = load_yaml("/opt/netbox/initializers/device_types.yml")
|
||||
|
||||
if device_types is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["manufacturer", "model", "slug"]
|
||||
required_assocs = {"manufacturer": (Manufacturer, "name")}
|
||||
optional_assocs = {"region": (Region, "name"), "tenant": (Tenant, "name")}
|
||||
nested_assocs = {"rear_port": (RearPortTemplate, "name"), "power_port": (PowerPortTemplate, "name")}
|
||||
|
||||
supported_components = {
|
||||
"interfaces": (InterfaceTemplate, ["name"]),
|
||||
"console_ports": (ConsolePortTemplate, ["name"]),
|
||||
"console_server_ports": (ConsoleServerPortTemplate, ["name"]),
|
||||
"power_ports": (PowerPortTemplate, ["name"]),
|
||||
"power_outlets": (PowerOutletTemplate, ["name"]),
|
||||
"rear_ports": (RearPortTemplate, ["name"]),
|
||||
"front_ports": (FrontPortTemplate, ["name"]),
|
||||
"device_bays": (DeviceBayTemplate, ["name"]),
|
||||
}
|
||||
|
||||
for params in device_types:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
components = [(v[0], v[1], params.pop(k, [])) for k, v in supported_components.items()]
|
||||
|
||||
for assoc, details in required_assocs.items():
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
for assoc, details in optional_assocs.items():
|
||||
if assoc in params:
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
device_type, created = DeviceType.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🔡 Created device type", device_type.manufacturer, device_type.model)
|
||||
|
||||
set_custom_fields_values(device_type, custom_field_data)
|
||||
|
||||
for component in components:
|
||||
c_model, c_match_params, c_params = component
|
||||
c_match_params.append("device_type")
|
||||
|
||||
if not c_params:
|
||||
continue
|
||||
|
||||
expanded_c_params = expand_templates(c_params, device_type)
|
||||
|
||||
for n_assoc, n_details in nested_assocs.items():
|
||||
n_model, n_field = n_details
|
||||
for c_param in expanded_c_params:
|
||||
if n_assoc in c_param:
|
||||
n_query = {n_field: c_param[n_assoc], "device_type": device_type}
|
||||
c_param[n_assoc] = n_model.objects.get(**n_query)
|
||||
|
||||
for new_param in expanded_c_params:
|
||||
new_matching_params, new_defaults = split_params(new_param, c_match_params)
|
||||
new_obj, new_obj_created = c_model.objects.get_or_create(
|
||||
**new_matching_params, defaults=new_defaults
|
||||
)
|
||||
if new_obj_created:
|
||||
print(
|
||||
f"🧷 Created {c_model._meta} {new_obj} component for device type {device_type}"
|
||||
)
|
@ -1,7 +1,12 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Device, DeviceRole, DeviceType, Platform, Rack, Site
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from dcim.models import Device, DeviceRole, DeviceType, Location, Platform, Rack, Site
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
from virtualization.models import Cluster
|
||||
|
||||
@ -10,23 +15,24 @@ devices = load_yaml("/opt/netbox/initializers/devices.yml")
|
||||
if devices is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["device_type", "name", "site"]
|
||||
required_assocs = {
|
||||
"device_role": (DeviceRole, "name"),
|
||||
"device_type": (DeviceType, "model"),
|
||||
"site": (Site, "name"),
|
||||
}
|
||||
|
||||
optional_assocs = {
|
||||
"tenant": (Tenant, "name"),
|
||||
"platform": (Platform, "name"),
|
||||
"rack": (Rack, "name"),
|
||||
"cluster": (Cluster, "name"),
|
||||
"location": (Location, "name"),
|
||||
}
|
||||
|
||||
for params in devices:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
|
||||
# primary ips are handled later in `270_primary_ips.py`
|
||||
# primary ips are handled later in `380_primary_ips.py`
|
||||
params.pop("primary_ip4", None)
|
||||
params.pop("primary_ip6", None)
|
||||
|
||||
@ -43,9 +49,10 @@ for params in devices:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
device, created = Device.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
device, created = Device.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(device, custom_field_data)
|
||||
|
||||
print("🖥️ Created device", device.name)
|
||||
|
||||
set_custom_fields_values(device, custom_field_data)
|
@ -1,29 +0,0 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Site
|
||||
from ipam.models import VLANGroup
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
|
||||
vlan_groups = load_yaml("/opt/netbox/initializers/vlan_groups.yml")
|
||||
|
||||
if vlan_groups is None:
|
||||
sys.exit()
|
||||
|
||||
optional_assocs = {"site": (Site, "name")}
|
||||
|
||||
for params in vlan_groups:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
|
||||
for assoc, details in optional_assocs.items():
|
||||
if assoc in params:
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
vlan_group, created = VLANGroup.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(vlan_group, custom_field_data)
|
||||
|
||||
print("🏘️ Created VLAN Group", vlan_group.name)
|
70
startup_scripts/210_dcim_interfaces.py
Normal file
70
startup_scripts/210_dcim_interfaces.py
Normal file
@ -0,0 +1,70 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Device, Interface
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
|
||||
interfaces = load_yaml("/opt/netbox/initializers/dcim_interfaces.yml")
|
||||
|
||||
if interfaces is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["device", "name"]
|
||||
required_assocs = {"device": (Device, "name")}
|
||||
related_assocs = {
|
||||
"bridge": (Interface, "name"),
|
||||
"lag": (Interface, "name"),
|
||||
"parent": (Interface, "name"),
|
||||
}
|
||||
|
||||
for params in interfaces:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
|
||||
related_interfaces = {k: params.pop(k, None) for k in related_assocs}
|
||||
|
||||
for assoc, details in required_assocs.items():
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
interface, created = Interface.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print(f"🧷 Created interface {interface} on {interface.device}")
|
||||
|
||||
set_custom_fields_values(interface, custom_field_data)
|
||||
|
||||
for related_field, related_value in related_interfaces.items():
|
||||
if not related_value:
|
||||
continue
|
||||
|
||||
r_model, r_field = related_assocs[related_field]
|
||||
|
||||
if related_field == "parent" and not interface.parent_id:
|
||||
query = {r_field: related_value, "device": interface.device}
|
||||
try:
|
||||
related_obj = r_model.objects.get(**query)
|
||||
except Interface.DoesNotExist:
|
||||
print(f"⚠️ Could not find parent interface with: {query} for interface {interface}")
|
||||
raise
|
||||
|
||||
interface.parent_id = related_obj.id
|
||||
interface.save()
|
||||
print(
|
||||
f"🧷 Attached interface {interface} on {interface.device} "
|
||||
f"to parent {related_obj}"
|
||||
)
|
||||
else:
|
||||
query = {r_field: related_value, "device": interface.device, "type": related_field}
|
||||
related_obj, rel_obj_created = r_model.objects.get_or_create(**query)
|
||||
|
||||
if rel_obj_created:
|
||||
setattr(interface, f"{related_field}_id", related_obj.id)
|
||||
interface.save()
|
||||
print(f"🧷 Created {related_field} interface {interface} on {interface.device}")
|
@ -1,7 +1,7 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Manufacturer, Platform
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
|
||||
platforms = load_yaml("/opt/netbox/initializers/platforms.yml")
|
||||
|
||||
@ -21,7 +21,8 @@ for params in platforms:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
platform, created = Platform.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
platform, created = Platform.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("💾 Created platform", platform.name)
|
@ -1,7 +1,12 @@
|
||||
import sys
|
||||
|
||||
from ipam.models import RouteTarget
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
|
||||
route_targets = load_yaml("/opt/netbox/initializers/route_targets.yml")
|
||||
@ -21,9 +26,10 @@ for params in route_targets:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
route_target, created = RouteTarget.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
route_target, created = RouteTarget.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(route_target, custom_field_data)
|
||||
|
||||
print("🎯 Created Route Target", route_target.name)
|
||||
|
||||
set_custom_fields_values(route_target, custom_field_data)
|
@ -1,7 +1,12 @@
|
||||
import sys
|
||||
|
||||
from ipam.models import VRF
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
|
||||
vrfs = load_yaml("/opt/netbox/initializers/vrfs.yml")
|
||||
@ -9,6 +14,7 @@ vrfs = load_yaml("/opt/netbox/initializers/vrfs.yml")
|
||||
if vrfs is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["name", "rd"]
|
||||
optional_assocs = {"tenant": (Tenant, "name")}
|
||||
|
||||
for params in vrfs:
|
||||
@ -21,9 +27,10 @@ for params in vrfs:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
vrf, created = VRF.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
vrf, created = VRF.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(vrf, custom_field_data)
|
||||
|
||||
print("📦 Created VRF", vrf.name)
|
||||
|
||||
set_custom_fields_values(vrf, custom_field_data)
|
@ -1,27 +0,0 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Device, Interface
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
|
||||
interfaces = load_yaml("/opt/netbox/initializers/dcim_interfaces.yml")
|
||||
|
||||
if interfaces is None:
|
||||
sys.exit()
|
||||
|
||||
required_assocs = {"device": (Device, "name")}
|
||||
|
||||
for params in interfaces:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
|
||||
for assoc, details in required_assocs.items():
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
interface, created = Interface.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(interface, custom_field_data)
|
||||
|
||||
print("🧷 Created interface", interface.name, interface.device.name)
|
@ -1,7 +1,7 @@
|
||||
import sys
|
||||
|
||||
from ipam.models import RIR
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
|
||||
rirs = load_yaml("/opt/netbox/initializers/rirs.yml")
|
||||
|
||||
@ -9,7 +9,8 @@ if rirs is None:
|
||||
sys.exit()
|
||||
|
||||
for params in rirs:
|
||||
rir, created = RIR.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
rir, created = RIR.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🗺️ Created RIR", rir.name)
|
34
startup_scripts/260_asns.py
Normal file
34
startup_scripts/260_asns.py
Normal file
@ -0,0 +1,34 @@
|
||||
import sys
|
||||
|
||||
from ipam.models import ASN, RIR
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
from tenancy.models import Tenant
|
||||
|
||||
asns = load_yaml("/opt/netbox/initializers/asns.yml")
|
||||
|
||||
if asns is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["asn", "rir"]
|
||||
required_assocs = {"rir": (RIR, "name")}
|
||||
optional_assocs = {"tenant": (Tenant, "name")}
|
||||
|
||||
for params in asns:
|
||||
for assoc, details in required_assocs.items():
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
for assoc, details in optional_assocs.items():
|
||||
if assoc in params:
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
asn, created = ASN.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print(f"🔡 Created ASN {asn.asn}")
|
@ -2,7 +2,12 @@ import sys
|
||||
|
||||
from ipam.models import RIR, Aggregate
|
||||
from netaddr import IPNetwork
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
|
||||
aggregates = load_yaml("/opt/netbox/initializers/aggregates.yml")
|
||||
@ -10,8 +15,8 @@ aggregates = load_yaml("/opt/netbox/initializers/aggregates.yml")
|
||||
if aggregates is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["prefix", "rir"]
|
||||
required_assocs = {"rir": (RIR, "name")}
|
||||
|
||||
optional_assocs = {
|
||||
"tenant": (Tenant, "name"),
|
||||
}
|
||||
@ -34,9 +39,10 @@ for params in aggregates:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
aggregate, created = Aggregate.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
aggregate, created = Aggregate.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(aggregate, custom_field_data)
|
||||
|
||||
print("🗞️ Created Aggregate", aggregate.prefix)
|
||||
|
||||
set_custom_fields_values(aggregate, custom_field_data)
|
@ -1,7 +1,7 @@
|
||||
import sys
|
||||
|
||||
from ipam.models import Role
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
|
||||
roles = load_yaml("/opt/netbox/initializers/prefix_vlan_roles.yml")
|
||||
|
||||
@ -9,7 +9,8 @@ if roles is None:
|
||||
sys.exit()
|
||||
|
||||
for params in roles:
|
||||
role, created = Role.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
role, created = Role.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("⛹️ Created Prefix/VLAN Role", role.name)
|
@ -1,19 +0,0 @@
|
||||
import sys
|
||||
|
||||
from circuits.models import Provider
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
|
||||
providers = load_yaml("/opt/netbox/initializers/providers.yml")
|
||||
|
||||
if providers is None:
|
||||
sys.exit()
|
||||
|
||||
for params in providers:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
|
||||
provider, created = Provider.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(provider, custom_field_data)
|
||||
|
||||
print("📡 Created provider", provider.name)
|
@ -1,6 +1,6 @@
|
||||
import sys
|
||||
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
from virtualization.models import ClusterType
|
||||
|
||||
cluster_types = load_yaml("/opt/netbox/initializers/cluster_types.yml")
|
||||
@ -9,7 +9,8 @@ if cluster_types is None:
|
||||
sys.exit()
|
||||
|
||||
for params in cluster_types:
|
||||
cluster_type, created = ClusterType.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
cluster_type, created = ClusterType.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🧰 Created Cluster Type", cluster_type.name)
|
@ -1,6 +1,6 @@
|
||||
import sys
|
||||
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
from virtualization.models import ClusterGroup
|
||||
|
||||
cluster_groups = load_yaml("/opt/netbox/initializers/cluster_groups.yml")
|
||||
@ -9,7 +9,10 @@ if cluster_groups is None:
|
||||
sys.exit()
|
||||
|
||||
for params in cluster_groups:
|
||||
cluster_group, created = ClusterGroup.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params)
|
||||
cluster_group, created = ClusterGroup.objects.get_or_create(
|
||||
**matching_params, defaults=defaults
|
||||
)
|
||||
|
||||
if created:
|
||||
print("🗄️ Created Cluster Group", cluster_group.name)
|
@ -1,7 +1,12 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import Site
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
from virtualization.models import Cluster, ClusterGroup, ClusterType
|
||||
|
||||
@ -10,8 +15,8 @@ clusters = load_yaml("/opt/netbox/initializers/clusters.yml")
|
||||
if clusters is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["name", "type"]
|
||||
required_assocs = {"type": (ClusterType, "name")}
|
||||
|
||||
optional_assocs = {
|
||||
"site": (Site, "name"),
|
||||
"group": (ClusterGroup, "name"),
|
||||
@ -34,9 +39,10 @@ for params in clusters:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
cluster, created = Cluster.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
cluster, created = Cluster.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(cluster, custom_field_data)
|
||||
|
||||
print("🗄️ Created cluster", cluster.name)
|
||||
|
||||
set_custom_fields_values(cluster, custom_field_data)
|
@ -1,15 +0,0 @@
|
||||
import sys
|
||||
from secrets.models import SecretRole
|
||||
|
||||
from startup_script_utils import load_yaml
|
||||
|
||||
secret_roles = load_yaml("/opt/netbox/initializers/secret_roles.yml")
|
||||
|
||||
if secret_roles is None:
|
||||
sys.exit()
|
||||
|
||||
for params in secret_roles:
|
||||
secret_role, created = SecretRole.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
||||
print("🔑 Created Secret Role", secret_role.name)
|
47
startup_scripts/320_vlan_groups.py
Normal file
47
startup_scripts/320_vlan_groups.py
Normal file
@ -0,0 +1,47 @@
|
||||
import sys
|
||||
|
||||
from django.contrib.contenttypes.models import ContentType
|
||||
from ipam.models import VLANGroup
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
|
||||
vlan_groups = load_yaml("/opt/netbox/initializers/vlan_groups.yml")
|
||||
|
||||
if vlan_groups is None:
|
||||
sys.exit()
|
||||
|
||||
optional_assocs = {"scope": (None, "name")}
|
||||
|
||||
for params in vlan_groups:
|
||||
custom_field_data = pop_custom_fields(params)
|
||||
|
||||
for assoc, details in optional_assocs.items():
|
||||
if assoc in params:
|
||||
model, field = details
|
||||
query = {field: params.pop(assoc)}
|
||||
# Get model from Contenttype
|
||||
scope_type = params.pop("scope_type", None)
|
||||
if not scope_type:
|
||||
print(f"VLAN Group '{params['name']}': scope_type is missing from VLAN Group")
|
||||
continue
|
||||
app_label, model = str(scope_type).split(".")
|
||||
ct = ContentType.objects.filter(app_label=app_label, model=model).first()
|
||||
if not ct:
|
||||
print(
|
||||
f"VLAN Group '{params['name']}': ContentType for "
|
||||
+ f"app_label = '{app_label}' and model = '{model}' not found"
|
||||
)
|
||||
continue
|
||||
params["scope_id"] = ct.model_class().objects.get(**query).id
|
||||
|
||||
matching_params, defaults = split_params(params)
|
||||
vlan_group, created = VLANGroup.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
print("🏘️ Created VLAN Group", vlan_group.name)
|
||||
|
||||
set_custom_fields_values(vlan_group, custom_field_data)
|
@ -2,7 +2,12 @@ import sys
|
||||
|
||||
from dcim.models import Site
|
||||
from ipam.models import VLAN, Role, VLANGroup
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant, TenantGroup
|
||||
|
||||
vlans = load_yaml("/opt/netbox/initializers/vlans.yml")
|
||||
@ -10,6 +15,7 @@ vlans = load_yaml("/opt/netbox/initializers/vlans.yml")
|
||||
if vlans is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["name", "vid"]
|
||||
optional_assocs = {
|
||||
"site": (Site, "name"),
|
||||
"tenant": (Tenant, "name"),
|
||||
@ -28,9 +34,10 @@ for params in vlans:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
vlan, created = VLAN.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
vlan, created = VLAN.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(vlan, custom_field_data)
|
||||
|
||||
print("🏠 Created VLAN", vlan.name)
|
||||
|
||||
set_custom_fields_values(vlan, custom_field_data)
|
@ -1,7 +1,12 @@
|
||||
import sys
|
||||
|
||||
from dcim.models import DeviceRole, Platform
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
from virtualization.models import Cluster, VirtualMachine
|
||||
|
||||
@ -10,8 +15,8 @@ virtual_machines = load_yaml("/opt/netbox/initializers/virtual_machines.yml")
|
||||
if virtual_machines is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["cluster", "name"]
|
||||
required_assocs = {"cluster": (Cluster, "name")}
|
||||
|
||||
optional_assocs = {
|
||||
"tenant": (Tenant, "name"),
|
||||
"platform": (Platform, "name"),
|
||||
@ -38,9 +43,12 @@ for params in virtual_machines:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
virtual_machine, created = VirtualMachine.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
virtual_machine, created = VirtualMachine.objects.get_or_create(
|
||||
**matching_params, defaults=defaults
|
||||
)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(virtual_machine, custom_field_data)
|
||||
|
||||
print("🖥️ Created virtual machine", virtual_machine.name)
|
||||
|
||||
set_custom_fields_values(virtual_machine, custom_field_data)
|
@ -1,6 +1,11 @@
|
||||
import sys
|
||||
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from virtualization.models import VirtualMachine, VMInterface
|
||||
|
||||
interfaces = load_yaml("/opt/netbox/initializers/virtualization_interfaces.yml")
|
||||
@ -8,6 +13,7 @@ interfaces = load_yaml("/opt/netbox/initializers/virtualization_interfaces.yml")
|
||||
if interfaces is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["name", "virtual_machine"]
|
||||
required_assocs = {"virtual_machine": (VirtualMachine, "name")}
|
||||
|
||||
for params in interfaces:
|
||||
@ -19,9 +25,10 @@ for params in interfaces:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
interface, created = VMInterface.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
interface, created = VMInterface.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(interface, custom_field_data)
|
||||
|
||||
print("🧷 Created interface", interface.name, interface.virtual_machine.name)
|
||||
|
||||
set_custom_fields_values(interface, custom_field_data)
|
@ -3,7 +3,12 @@ import sys
|
||||
from dcim.models import Site
|
||||
from ipam.models import VLAN, VRF, Prefix, Role
|
||||
from netaddr import IPNetwork
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant, TenantGroup
|
||||
|
||||
prefixes = load_yaml("/opt/netbox/initializers/prefixes.yml")
|
||||
@ -11,6 +16,7 @@ prefixes = load_yaml("/opt/netbox/initializers/prefixes.yml")
|
||||
if prefixes is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["prefix", "site", "vrf", "vlan"]
|
||||
optional_assocs = {
|
||||
"site": (Site, "name"),
|
||||
"tenant": (Tenant, "name"),
|
||||
@ -31,9 +37,10 @@ for params in prefixes:
|
||||
query = {field: params.pop(assoc)}
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
prefix, created = Prefix.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
prefix, created = Prefix.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(prefix, custom_field_data)
|
||||
|
||||
print("📌 Created Prefix", prefix.prefix)
|
||||
|
||||
set_custom_fields_values(prefix, custom_field_data)
|
@ -5,7 +5,12 @@ from django.contrib.contenttypes.models import ContentType
|
||||
from django.db.models import Q
|
||||
from ipam.models import VRF, IPAddress
|
||||
from netaddr import IPNetwork
|
||||
from startup_script_utils import load_yaml, pop_custom_fields, set_custom_fields_values
|
||||
from startup_script_utils import (
|
||||
load_yaml,
|
||||
pop_custom_fields,
|
||||
set_custom_fields_values,
|
||||
split_params,
|
||||
)
|
||||
from tenancy.models import Tenant
|
||||
from virtualization.models import VirtualMachine, VMInterface
|
||||
|
||||
@ -14,10 +19,11 @@ ip_addresses = load_yaml("/opt/netbox/initializers/ip_addresses.yml")
|
||||
if ip_addresses is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["address", "vrf"]
|
||||
optional_assocs = {
|
||||
"tenant": (Tenant, "name"),
|
||||
"vrf": (VRF, "name"),
|
||||
"interface": (None, None),
|
||||
"interface": (Interface, "name"),
|
||||
}
|
||||
|
||||
vm_interface_ct = ContentType.objects.filter(
|
||||
@ -52,11 +58,13 @@ for params in ip_addresses:
|
||||
params["assigned_object_id"] = Interface.objects.get(**query).id
|
||||
else:
|
||||
query = {field: params.pop(assoc)}
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
ip_address, created = IPAddress.objects.get_or_create(**params)
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
ip_address, created = IPAddress.objects.get_or_create(**matching_params, defaults=defaults)
|
||||
|
||||
if created:
|
||||
set_custom_fields_values(ip_address, custom_field_data)
|
||||
|
||||
print("🧬 Created IP Address", ip_address.address)
|
||||
|
||||
set_custom_fields_values(ip_address, custom_field_data)
|
@ -2,7 +2,7 @@ import sys
|
||||
|
||||
from dcim.models import Device
|
||||
from ipam.models import Service
|
||||
from startup_script_utils import load_yaml
|
||||
from startup_script_utils import load_yaml, split_params
|
||||
from virtualization.models import VirtualMachine
|
||||
|
||||
services = load_yaml("/opt/netbox/initializers/services.yml")
|
||||
@ -10,6 +10,7 @@ services = load_yaml("/opt/netbox/initializers/services.yml")
|
||||
if services is None:
|
||||
sys.exit()
|
||||
|
||||
match_params = ["name", "device", "virtual_machine"]
|
||||
optional_assocs = {
|
||||
"device": (Device, "name"),
|
||||
"virtual_machine": (VirtualMachine, "name"),
|
||||
@ -24,6 +25,7 @@ for params in services:
|
||||
|
||||
params[assoc] = model.objects.get(**query)
|
||||
|
||||
matching_params, defaults = split_params(params, match_params)
|
||||
service, created = Service.objects.get_or_create(**params)
|
||||
|
||||
if created:
|
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user