Compare commits
913 Commits
Author | SHA1 | Date | |
---|---|---|---|
bd00ff10e4 | |||
149b969d9a | |||
56d3b59c71 | |||
c1e6efa8e1 | |||
3b5473a682 | |||
4954d3130b | |||
064497756e | |||
ce3c7a1bda | |||
50a39bbc1f | |||
154d01b042 | |||
1f3352018b | |||
b721783c48 | |||
76ee3085a4 | |||
5d5a53059f | |||
77d8c593b3 | |||
c450a3cafd | |||
f8f4d7cab4 | |||
91abfef049 | |||
963b7ec51b | |||
16aab0c137 | |||
bf8b8be976 | |||
e201104d0b | |||
d63db0863d | |||
7a36833103 | |||
ca6e66aa5a | |||
94a6b33680 | |||
2d5287fbbc | |||
6eb756bcab | |||
5647219049 | |||
b810972823 | |||
3a07cdf574 | |||
193ec30c2b | |||
c94723062c | |||
0eadfdf670 | |||
ebf8ce20bc | |||
e7acdde758 | |||
f2c9da2349 | |||
ff344655e2 | |||
3490d9460c | |||
fdf9373f9e | |||
ba80611324 | |||
07a579c632 | |||
188a37fbed | |||
f251367c33 | |||
ac4e399a10 | |||
4fe77c36df | |||
118515dbd0 | |||
42ba4cd399 | |||
ab1c07a622 | |||
930a71460f | |||
a58a5cf795 | |||
92a8f0bc82 | |||
bf298a16ef | |||
9a1b24b6b1 | |||
ea67cd70c9 | |||
281a5dd1fc | |||
df3b3d1798 | |||
f5e2b4726d | |||
daaeea8b4b | |||
6f6df501a0 | |||
d5790a9f27 | |||
860eaec58f | |||
26e949d5fe | |||
ac7dbba458 | |||
7c2431d42c | |||
25c1420a12 | |||
c1a1e1ae8f | |||
a9df9df25d | |||
10beed1199 | |||
df32530750 | |||
062edce27f | |||
efd2713aa8 | |||
8a21566c8a | |||
c8c5c7f571 | |||
91357c2034 | |||
097ccfe1d5 | |||
61ef4ae8cb | |||
01ae7bfaf2 | |||
1b52122a1f | |||
1d9bc184f5 | |||
5f83d3f636 | |||
71e534631f | |||
6e9e6c7a54 | |||
e2e7560d5e | |||
0ceb97538a | |||
3e276f6fb6 | |||
2b00c5abca | |||
15cc41b6cb | |||
729bd1fd16 | |||
9a7431e2e0 | |||
52fbc86fc9 | |||
afe6c79ce3 | |||
9407810fe1 | |||
c42a54795d | |||
96ec3801a9 | |||
c4707d0c1d | |||
24f9af9e0f | |||
a0172d766b | |||
09f999337a | |||
e3eb062c09 | |||
de21d4efdc | |||
d5f58006d3 | |||
cb80ffc1de | |||
1859a0eb8b | |||
9e7132c0b3 | |||
bf013be1c4 | |||
b935209584 | |||
efd4ddc17b | |||
e511e0e553 | |||
610150a4b4 | |||
485b2438ac | |||
bfd12e871f | |||
0c136bfab1 | |||
245e2aea23 | |||
b9d588ffde | |||
e4bc3e0e8d | |||
2419dc0de9 | |||
68fd9ca6d6 | |||
4beb7d2dbe | |||
2bc1250c28 | |||
9b1e2ae83c | |||
9b5ecbe2ff | |||
342ed4aea0 | |||
d4e9d5470e | |||
5c1cabdea1 | |||
38517ca053 | |||
e33758d1b8 | |||
aba6189c4f | |||
adcc21716b | |||
87e17fb4d1 | |||
8292d3d20e | |||
5cc7d89139 | |||
343392613d | |||
de91418b79 | |||
fe9c47ab4f | |||
02db72678f | |||
db4b469285 | |||
92c5cf42d1 | |||
e9558f290a | |||
572e6594d2 | |||
88691284d8 | |||
85c622807e | |||
9d42e0475b | |||
181a335bfa | |||
0a33951e9e | |||
7a356a748a | |||
1c402740a2 | |||
e0a19d3313 | |||
6b8329ee34 | |||
1d4448998a | |||
d6473f5359 | |||
f5f9ec81d2 | |||
fea950155f | |||
ef2944bc24 | |||
934c8724e2 | |||
98eb435d90 | |||
bd10af6eda | |||
7f381a6246 | |||
c17fbbbc07 | |||
ac2ca6c341 | |||
d26865c52c | |||
2b05008a11 | |||
45700e2ecf | |||
f84304235b | |||
0ca41155b2 | |||
a291ab59ba | |||
fce7cd0d36 | |||
658357c5a8 | |||
7484fce24d | |||
f28a713e2b | |||
a9017805b7 | |||
2e3f94e12f | |||
d531c7ae61 | |||
7df1580fa6 | |||
58f70bccbb | |||
fae4f6c509 | |||
ddafb28572 | |||
642c7b9915 | |||
5a8726e6d2 | |||
b3f279e2d9 | |||
82f5ad18f0 | |||
bacc99c7f8 | |||
6728d0977b | |||
bff7c027c9 | |||
79b3113361 | |||
5885767b91 | |||
ec08247e5c | |||
400f081487 | |||
03664514ab | |||
c68fa58a59 | |||
426dda0730 | |||
eb37d4ece2 | |||
1198f8d4e6 | |||
4b709ade68 | |||
fa49d0fde9 | |||
1d44f175c6 | |||
890b88cbef | |||
27709b49d5 | |||
7ccbce03d3 | |||
5fb852afed | |||
60589e6066 | |||
717ce40612 | |||
75442e813e | |||
853c55a049 | |||
6ef1b649d9 | |||
e3f3359c86 | |||
0e1edf19b1 | |||
de55fff226 | |||
b3a67f1f14 | |||
3cc23ca6cc | |||
3def6bfc64 | |||
18e8bc17e4 | |||
f66d66aafe | |||
7380c48dff | |||
0191759316 | |||
dbc42e6f75 | |||
d1c3bc5350 | |||
a97301350f | |||
09340f28f5 | |||
20497c6346 | |||
d0f7d0d9c1 | |||
608806e884 | |||
48176b0a77 | |||
3483a3b3a1 | |||
347e0d4c57 | |||
ae9b5c077a | |||
747446eb50 | |||
e1c8c27f47 | |||
63cec1622a | |||
31142ef291 | |||
058b4b9708 | |||
9a1330c72e | |||
0a6df20986 | |||
6680878b5c | |||
593043ed53 | |||
038f385089 | |||
b914b94773 | |||
2194bc59c8 | |||
a98a288e2d | |||
49e25688f1 | |||
d7eedbd24b | |||
5b17a02da4 | |||
8735247f29 | |||
0d5d15c9d1 | |||
2e44983a37 | |||
c76ff4b472 | |||
aaf4f40285 | |||
e64f77b716 | |||
fd1b65cc3c | |||
11148dce43 | |||
38da8ca1bc | |||
a0ffd4a413 | |||
450105b0c3 | |||
b62edce929 | |||
67678ec39c | |||
bf95fba72e | |||
d265420025 | |||
01a080215d | |||
8cf445ecc4 | |||
20def38e96 | |||
be5b43cb87 | |||
6f0565fa60 | |||
99940358e3 | |||
53daae8e89 | |||
8a23ea4656 | |||
c95c1c83b0 | |||
b446fa14c5 | |||
6d5d305d9d | |||
af2eb422d5 | |||
bbd57396d7 | |||
0fd55b08d9 | |||
619cd5cbcb | |||
1ec0d70d09 | |||
c8449217dc | |||
f7348a23cd | |||
ae18c436dd | |||
b0e20a71e2 | |||
b9700a9fe5 | |||
81867f0539 | |||
0a33fba49c | |||
049a22a3a3 | |||
4d4f94dedf | |||
a844fa0ba0 | |||
497a7b3f8e | |||
71549afa3f | |||
a294588409 | |||
5a83930667 | |||
c25ea25f0a | |||
f7885eb263 | |||
a48d534d39 | |||
bfa942c0cf | |||
f54634a890 | |||
efb7c5348c | |||
d6fcc1170a | |||
3f742f952a | |||
84af82e8cf | |||
48109c5354 | |||
fd18775ac1 | |||
e678a50ea1 | |||
6523588c8d | |||
6fbf0acc76 | |||
36b7085ec2 | |||
1b1a553741 | |||
98b7d58b94 | |||
7fa9a37c7c | |||
f533d16ef6 | |||
778c7d954b | |||
605fe2e7e7 | |||
1b552c109d | |||
d4d49f7325 | |||
8bca935f08 | |||
fd6d243843 | |||
037f6b6d5e | |||
8eef31724f | |||
2de1b06a06 | |||
a332040a7f | |||
957133077f | |||
36c6e7bb82 | |||
ccc3896ff3 | |||
cef5c72682 | |||
51a2d9e375 | |||
048b43af24 | |||
bfd2b47649 | |||
67a5cf4714 | |||
6227654ad8 | |||
e384f16a19 | |||
89725197c0 | |||
e7d4be9d85 | |||
ba3d7e19fb | |||
b65dfff574 | |||
8cc3760e74 | |||
1cb08a0a05 | |||
6f4228809e | |||
5af3bcf062 | |||
67d00d5c0e | |||
cdc83c4eb2 | |||
ffa403b5fd | |||
5bd77f00e2 | |||
802189f7f5 | |||
a4e5a0fc9f | |||
58bfa3b19c | |||
f9c0a94140 | |||
e3619d4101 | |||
5839c469c1 | |||
bbdda58b35 | |||
ed2080762c | |||
45d5d873ce | |||
f46806414a | |||
ebf34e7edd | |||
aad2d162ab | |||
68149b9045 | |||
1ce8e905ea | |||
ccb3b45e18 | |||
6afdda8832 | |||
2121174827 | |||
df12c9ec4e | |||
4c1b776168 | |||
42dad3abd3 | |||
6c76aa434d | |||
e5f9b7f79e | |||
dd2162f6bd | |||
cabdabba3d | |||
3e593a2459 | |||
7c5287bb95 | |||
7c72ae04f1 | |||
86582454e8 | |||
013b1e8bca | |||
40ff84b138 | |||
b2065dc7d2 | |||
97dfc62f0d | |||
e351ac786d | |||
7b570c177d | |||
6838b75904 | |||
dbda1513c5 | |||
c62a6acb2e | |||
e4a5c072b4 | |||
80f950c05d | |||
4933b853cd | |||
aec1b91eb8 | |||
2e2d64fdba | |||
a37c8d2431 | |||
a8a20e9210 | |||
be5b468975 | |||
9789461363 | |||
9f58e312d7 | |||
cffe0b81e3 | |||
bb14ed8cab | |||
023adb5945 | |||
e5545c9804 | |||
efe96ec039 | |||
1d3ae83359 | |||
4bb3876352 | |||
400e90cfbe | |||
e16c289f50 | |||
140c159b36 | |||
8be69a8453 | |||
9ba4833f3c | |||
0b12a5a698 | |||
2eac359430 | |||
855b55dc14 | |||
5ad40a3dd1 | |||
7116a2d9da | |||
0d5e990a62 | |||
4f57f4ad84 | |||
13e13d836f | |||
3ab2432ab6 | |||
76e8565076 | |||
a5f30a562b | |||
a2ef36d445 | |||
9a1ecae0b7 | |||
42b010174e | |||
68e77657e6 | |||
1b2f851e42 | |||
cc99866ea3 | |||
1ea3f23f7e | |||
3f780ddf73 | |||
9edf96e6b6 | |||
73e1ba65ca | |||
02631056b8 | |||
131d0f10c2 | |||
f9aa980c7d | |||
ad0364c558 | |||
76486eb3d1 | |||
65ab4ca976 | |||
99a73fad15 | |||
16a01c19dd | |||
86b8ba448c | |||
9b8e8012a7 | |||
b29292a87b | |||
c1feb447e8 | |||
62a0e190cb | |||
5890143920 | |||
ef4df211ab | |||
eb5e0ae65a | |||
bbc71e3b02 | |||
ac81ed17b9 | |||
89145cde34 | |||
ef4b2c2470 | |||
7190cbf2ac | |||
f726e1e0ea | |||
6d81e65986 | |||
ba5f5083c3 | |||
314db4072c | |||
baff2324f3 | |||
02eae829f7 | |||
bb77143108 | |||
02cb5b5f80 | |||
a301c362e3 | |||
7526d86419 | |||
a00888e93f | |||
fc5870be53 | |||
3c8c2827cb | |||
6c221244df | |||
38629c3961 | |||
513d019ac3 | |||
3fa1b4b48c | |||
a6eac535e4 | |||
58a3fae773 | |||
0889806a3c | |||
51ec8a3c62 | |||
a12b1be728 | |||
4d04cd9ab9 | |||
a3399f4337 | |||
2b7f8dd5ea | |||
72fbe9ffa5 | |||
0be8bce718 | |||
4805edc4ec | |||
9eb784076c | |||
b9c5cd8291 | |||
9008c0c177 | |||
f027c2146e | |||
afbf2e10f3 | |||
9805207aa5 | |||
8e0b852f24 | |||
0052dc6d28 | |||
61f05679d2 | |||
9751ef4b36 | |||
0a240aaa9a | |||
e0665a64bd | |||
dc46aa9a00 | |||
ced694589d | |||
6c053ffc89 | |||
9f5b57a348 | |||
f1c4b8df34 | |||
269e274bb5 | |||
bfd357c5a1 | |||
9517a5759a | |||
a5d51b0c4f | |||
d9822cd3cb | |||
66501529a2 | |||
2072dede4a | |||
31c94d1645 | |||
9ee4c23833 | |||
a14a1c7b90 | |||
9ef88578af | |||
c4c4b5a3ef | |||
0ed40b19c7 | |||
a0cd0f9cec | |||
49e47c491b | |||
424d2d68d3 | |||
415690a0e7 | |||
2c0abe9234 | |||
2649c89358 | |||
bbd34d70d5 | |||
9779ad0b00 | |||
70fd0652a1 | |||
6b85671dd2 | |||
82bdf6b5e7 | |||
ba2679c9d7 | |||
8866cbccc8 | |||
b3477d286f | |||
68e2ea99ba | |||
d6688884f6 | |||
7d3482f5bf | |||
7a39b41c20 | |||
4672273fe6 | |||
01284de0b2 | |||
b20368ee1b | |||
e584593cb5 | |||
069a6e28a7 | |||
8fab19da73 | |||
991be99c37 | |||
1900d7810c | |||
6b5013edb3 | |||
f313494d48 | |||
353dcf1d13 | |||
3006d70ebe | |||
681e096448 | |||
ac9a9e8002 | |||
ecbc385b7b | |||
5117cf4f17 | |||
da7ec1d2af | |||
934de1d691 | |||
0c27d880b0 | |||
be3a0295b6 | |||
aa2838c27a | |||
ea584a7510 | |||
ba0ccc5991 | |||
75f83c6a81 | |||
0dda5a6695 | |||
289738dc1a | |||
d830804f02 | |||
82cc4b56e5 | |||
923f94a4d7 | |||
bbff317aa7 | |||
20429238e0 | |||
364299740f | |||
b81818b6ad | |||
2f02e431b0 | |||
e64f38cb6b | |||
ae24382634 | |||
82cae19d19 | |||
3f5fbc5620 | |||
000e6cad5c | |||
49f44cedbf | |||
eb1c59cc2a | |||
c7d032fc17 | |||
73b77d4787 | |||
67466ce564 | |||
4e0faf5ef3 | |||
c23192d34e | |||
83771aa037 | |||
95f9d67ce9 | |||
314d360fcd | |||
f8a74456cc | |||
4906bac10f | |||
86c831a5c3 | |||
a5951b4f38 | |||
f75292bd8d | |||
bfff4eaa7f | |||
067dc06dba | |||
18cdf20afc | |||
e57841c442 | |||
751f6b6148 | |||
3c430e9a55 | |||
155f657f6b | |||
86fb38776b | |||
f323e90602 | |||
770a36e53a | |||
d420962fbc | |||
01fd2447b2 | |||
85beb7d875 | |||
af06decd1b | |||
aceae32baa | |||
74a4f9efc9 | |||
fb1e7a86f4 | |||
dc99315cf9 | |||
34bd1109b0 | |||
13a2445744 | |||
c968da789e | |||
3f84541412 | |||
4d8bd03668 | |||
f9bd5e1691 | |||
ecd66ecaf6 | |||
33d7292f29 | |||
f4d371d2d2 | |||
835d0e5dd3 | |||
9a06eb1618 | |||
309e14ebb7 | |||
2d48533378 | |||
fffd6874e6 | |||
0ddd48f0b5 | |||
cb590dbc07 | |||
6c4f762c49 | |||
7a0afee391 | |||
0dda883994 | |||
c2e2078b3f | |||
26a3450f19 | |||
324c069848 | |||
bd4c5607ca | |||
e1d85f1840 | |||
1ce1a5e5cc | |||
6f66a0ca71 | |||
62a5b3907b | |||
85b6c4ead4 | |||
a190979c04 | |||
4a489ae3de | |||
9ac8b73e07 | |||
414be8b675 | |||
fda19dcc6f | |||
cd975e5787 | |||
3b7b1dfb8e | |||
d8a47ec649 | |||
252cd3b781 | |||
0decd11efb | |||
b84d2592fb | |||
0219ba2cc5 | |||
bbff6c4968 | |||
bb88c6a29d | |||
a02466966d | |||
b0fc11804e | |||
d9d81741e3 | |||
9678366102 | |||
a2c73c78dd | |||
c6a0e7d98e | |||
85417b2a88 | |||
d738669066 | |||
442d6da8fb | |||
62f10a01db | |||
5667b76381 | |||
d9b318a444 | |||
86ce56f193 | |||
8d72c2c32e | |||
c48c38ab8c | |||
3d3769830b | |||
4921a411ad | |||
81c767efce | |||
60abf03f05 | |||
dcbf29e71b | |||
037e6c0ca8 | |||
c7024b282a | |||
90ff75f85c | |||
2165f0d450 | |||
1e7639bfc4 | |||
4121628d99 | |||
da78b90f9c | |||
1ef6e8b6a7 | |||
10351f7075 | |||
70a152deb7 | |||
5446bfbba8 | |||
400885e620 | |||
f960fc3b6f | |||
ddfa4d679a | |||
10e8026786 | |||
2527c039df | |||
93d8a2044e | |||
d2354a16cd | |||
34ee1f1c76 | |||
2de4dc3a81 | |||
b90036dadd | |||
4708f4fc21 | |||
062cf75cdf | |||
e5950360ca | |||
5b358ff0b1 | |||
4c00391d78 | |||
9594362e35 | |||
3420029b5e | |||
f432a1c927 | |||
e8b32f2d87 | |||
3e3b505cc8 | |||
0bca966ec5 | |||
84737fb33f | |||
e21a15ab17 | |||
90066d22a0 | |||
dbf5dad1c4 | |||
c793da1edc | |||
f8735e5988 | |||
e9805b2486 | |||
eb90405a78 | |||
ecf5f468c3 | |||
51aee8cac8 | |||
7d5049c350 | |||
01a99f5651 | |||
2914e99ff3 | |||
f9b824ac30 | |||
9a535ec77b | |||
ffba023c91 | |||
e01689978e | |||
68ac8976eb | |||
afb790db73 | |||
0732de361a | |||
d455270fa1 | |||
1336be16c9 | |||
03380db560 | |||
927ebc702c | |||
c24cb13382 | |||
3a804a8a20 | |||
1fde4167ea | |||
75f9f40922 | |||
e9c2638f90 | |||
338c545f85 | |||
e379b4a31c | |||
3d7ca2bdb9 | |||
d34019e246 | |||
7cb2ebba79 | |||
4e8581950e | |||
2a9a3d632e | |||
b6d07fa038 | |||
4599e7959c | |||
82ed13c7d7 | |||
5aaa81ab89 | |||
8a06d1935e | |||
f44254b4bd | |||
07875ce13e | |||
98dc770efa | |||
8848f1d487 | |||
5128ae48a0 | |||
104ae6093a | |||
e830d63f6a | |||
ce32cd487a | |||
f36c659365 | |||
47e5cbdb03 | |||
4923a76f22 | |||
e01ca6a2dd | |||
5e989333cd | |||
af39c399bc | |||
64591e731e | |||
5658504b90 | |||
64e0786aa9 | |||
90761f0f62 | |||
74f74d1e64 | |||
4db4b9706c | |||
00a5072ad3 | |||
3d3d698bb3 | |||
1b9521bb87 | |||
1d781c5b20 | |||
8e8836d1ea | |||
a904e3755d | |||
7ba99fef86 | |||
7d2be91bc9 | |||
578895336a | |||
8c090937f5 | |||
4229633d98 | |||
3ed7e87538 | |||
5b43cc4487 | |||
3241392117 | |||
c474a66b41 | |||
b32cf6a1e0 | |||
f32791b4b2 | |||
8f33fe8e59 | |||
d19010481d | |||
6b11524a8b | |||
e953029e8f | |||
10f788b7eb | |||
9348544e46 | |||
126ccbcfa6 | |||
440472cb32 | |||
4ce7da516d | |||
a7f8efcf35 | |||
9fe4c79005 | |||
f09f4d5fd5 | |||
38b4f9b534 | |||
fca1cef29f | |||
45b8a0327f | |||
a723c08715 | |||
c381a162fb | |||
b4931192c3 | |||
cc269b9ff9 | |||
a5e3be4992 | |||
137309cc4e | |||
85f4e834d8 | |||
065013ccec | |||
56d98ba966 | |||
dda1b4fa44 | |||
68b102269f | |||
0ecdaa0dc0 | |||
13f435caab | |||
ff99780303 | |||
fa9507020a | |||
1bff50afea | |||
37ff72720b | |||
2d5d264f99 | |||
c9c07445b7 | |||
a4388ffc36 | |||
ea1458923e | |||
e857f1fae8 | |||
3ec42e81b1 | |||
be1163acfe | |||
d308dc8af7 | |||
60643023ad | |||
875d53ef6c | |||
b41f9e9fec | |||
a1b71c3c7d | |||
013fa2d886 | |||
72e311c6b2 | |||
2732c47466 | |||
0466089316 | |||
5e42d38598 | |||
82a4bb5e80 | |||
94bc7957c1 | |||
c9e6b07145 | |||
3c06eba17a | |||
8081e4aa7b | |||
d8769d659e | |||
572cd0381b | |||
5e91b40087 | |||
936eceda61 | |||
61c4087041 | |||
7d39e47182 | |||
c4e1af3069 | |||
3e234af16e | |||
bbbf662d20 | |||
25d78b1068 | |||
78bf292343 | |||
e5ef69ecf7 | |||
b7b9a57425 | |||
c4a04b7c62 | |||
2e41dbe828 | |||
56d36ca439 | |||
e0ba5553be | |||
8d6fb677c1 | |||
a2daecc25d | |||
ee0c5c8e01 | |||
ae5b1e188f | |||
49f9aca627 | |||
4cba875379 | |||
7ab4382476 | |||
eaef6c8d00 | |||
95f3692545 | |||
686173dc2a | |||
39c5db7f0f | |||
603aa09d54 | |||
88aa3076f0 | |||
5400fe171c | |||
87bf9f569f | |||
8fb24a2c0a | |||
4b5d9b6e64 | |||
72bd8293e3 | |||
09989d9963 | |||
4088d5bc62 | |||
d4b84c1dec | |||
426847e1ce | |||
79b902d512 | |||
73c607497e | |||
f2f526b61d | |||
cb67ecaddb | |||
5bf9b0b0bb | |||
7a61f89e5a | |||
671c6a96e7 | |||
f0d23e5370 | |||
d1bee4344d | |||
d724116c0c | |||
888d89e2dd | |||
a6471bc346 | |||
6b1da1c166 | |||
18210d8958 | |||
bc5c1a9aa6 | |||
3df77ef5da | |||
e8d9d9adfa | |||
01d152720f | |||
5e58381ea9 | |||
0b6d9442bd | |||
134ed9e14f | |||
0796b642de | |||
f912ba6a3e | |||
a576e6685b | |||
b1c793cfa5 | |||
c0147e49c4 | |||
d52b120905 | |||
84c8a580b5 | |||
467bd01cdf | |||
7a7fcb4715 | |||
cf8e44bc30 | |||
279e7eb497 | |||
606828cc65 | |||
aac424674c | |||
8fd1e10830 | |||
12509a6d9e | |||
5e169f387c | |||
8369ade880 | |||
73cef112eb | |||
4a0132382a | |||
6ee69fccd3 | |||
a862835be2 | |||
ddbd63ed5f | |||
6a59fa0e18 | |||
1ed9069ad3 | |||
a588b67906 | |||
37a634f550 | |||
951fe0cb7d | |||
4ca3f0c6ae | |||
69e5ba29c4 | |||
e045d154e9 | |||
6526709d48 | |||
603f80d813 | |||
398636b61c | |||
eb70464839 | |||
75054859ff | |||
8e898895cc | |||
4be6beab6f | |||
a3b4b5b50e | |||
33b8d7e5e8 | |||
f2f43e1904 |
83
Cargo.toml
83
Cargo.toml
@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "proxmox-backup"
|
name = "proxmox-backup"
|
||||||
version = "1.1.3"
|
version = "2.1.2"
|
||||||
authors = [
|
authors = [
|
||||||
"Dietmar Maurer <dietmar@proxmox.com>",
|
"Dietmar Maurer <dietmar@proxmox.com>",
|
||||||
"Dominik Csapak <d.csapak@proxmox.com>",
|
"Dominik Csapak <d.csapak@proxmox.com>",
|
||||||
@ -18,15 +18,35 @@ homepage = "https://www.proxmox.com"
|
|||||||
|
|
||||||
exclude = [ "build", "debian", "tests/catar_data/test_symlink/symlink1"]
|
exclude = [ "build", "debian", "tests/catar_data/test_symlink/symlink1"]
|
||||||
|
|
||||||
|
[workspace]
|
||||||
|
members = [
|
||||||
|
"pbs-buildcfg",
|
||||||
|
"pbs-client",
|
||||||
|
"pbs-config",
|
||||||
|
"pbs-datastore",
|
||||||
|
"pbs-fuse-loop",
|
||||||
|
"proxmox-rest-server",
|
||||||
|
"proxmox-rrd",
|
||||||
|
"pbs-tape",
|
||||||
|
"pbs-tools",
|
||||||
|
|
||||||
|
"proxmox-backup-banner",
|
||||||
|
"proxmox-backup-client",
|
||||||
|
"proxmox-file-restore",
|
||||||
|
"proxmox-restore-daemon",
|
||||||
|
"pxar-bin",
|
||||||
|
]
|
||||||
|
|
||||||
[lib]
|
[lib]
|
||||||
name = "proxmox_backup"
|
name = "proxmox_backup"
|
||||||
path = "src/lib.rs"
|
path = "src/lib.rs"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
apt-pkg-native = "0.3.2"
|
apt-pkg-native = "0.3.2"
|
||||||
base64 = "0.12"
|
base64 = "0.13"
|
||||||
bitflags = "1.2.1"
|
bitflags = "1.2.1"
|
||||||
bytes = "1.0"
|
bytes = "1.0"
|
||||||
|
cidr = "0.2.1"
|
||||||
crc32fast = "1"
|
crc32fast = "1"
|
||||||
endian_trait = { version = "0.6", features = ["arrays"] }
|
endian_trait = { version = "0.6", features = ["arrays"] }
|
||||||
env_logger = "0.7"
|
env_logger = "0.7"
|
||||||
@ -36,6 +56,7 @@ thiserror = "1.0"
|
|||||||
futures = "0.3"
|
futures = "0.3"
|
||||||
h2 = { version = "0.3", features = [ "stream" ] }
|
h2 = { version = "0.3", features = [ "stream" ] }
|
||||||
handlebars = "3.0"
|
handlebars = "3.0"
|
||||||
|
hex = "0.4.3"
|
||||||
http = "0.2"
|
http = "0.2"
|
||||||
hyper = { version = "0.14", features = [ "full" ] }
|
hyper = { version = "0.14", features = [ "full" ] }
|
||||||
lazy_static = "1.4"
|
lazy_static = "1.4"
|
||||||
@ -48,36 +69,68 @@ openssl = "0.10"
|
|||||||
pam = "0.7"
|
pam = "0.7"
|
||||||
pam-sys = "0.5"
|
pam-sys = "0.5"
|
||||||
percent-encoding = "2.1"
|
percent-encoding = "2.1"
|
||||||
pin-utils = "0.1.0"
|
|
||||||
pin-project = "1.0"
|
|
||||||
pathpatterns = "0.1.2"
|
|
||||||
proxmox = { version = "0.11.1", features = [ "sortable-macro", "api-macro", "websocket" ] }
|
|
||||||
#proxmox = { git = "git://git.proxmox.com/git/proxmox", version = "0.1.2", features = [ "sortable-macro", "api-macro" ] }
|
|
||||||
#proxmox = { path = "../proxmox/proxmox", features = [ "sortable-macro", "api-macro", "websocket" ] }
|
|
||||||
proxmox-fuse = "0.1.1"
|
|
||||||
pxar = { version = "0.10.1", features = [ "tokio-io" ] }
|
|
||||||
#pxar = { path = "../pxar", features = [ "tokio-io" ] }
|
|
||||||
regex = "1.2"
|
regex = "1.2"
|
||||||
rustyline = "7"
|
rustyline = "7"
|
||||||
serde = { version = "1.0", features = ["derive"] }
|
serde = { version = "1.0", features = ["derive"] }
|
||||||
serde_json = "1.0"
|
serde_json = "1.0"
|
||||||
siphasher = "0.3"
|
siphasher = "0.3"
|
||||||
syslog = "4.0"
|
syslog = "4.0"
|
||||||
tokio = { version = "1.0", features = [ "fs", "io-util", "io-std", "macros", "net", "parking_lot", "process", "rt", "rt-multi-thread", "signal", "time" ] }
|
tokio = { version = "1.6", features = [ "fs", "io-util", "io-std", "macros", "net", "parking_lot", "process", "rt", "rt-multi-thread", "signal", "time" ] }
|
||||||
tokio-openssl = "0.6.1"
|
tokio-openssl = "0.6.1"
|
||||||
tokio-stream = "0.1.0"
|
tokio-stream = "0.1.0"
|
||||||
tokio-util = { version = "0.6", features = [ "codec", "io" ] }
|
tokio-util = { version = "0.6", features = [ "codec", "io" ] }
|
||||||
tower-service = "0.3.0"
|
tower-service = "0.3.0"
|
||||||
udev = ">= 0.3, <0.5"
|
udev = "0.4"
|
||||||
url = "2.1"
|
url = "2.1"
|
||||||
#valgrind_request = { git = "https://github.com/edef1c/libvalgrind_request", version = "1.1.0", optional = true }
|
#valgrind_request = { git = "https://github.com/edef1c/libvalgrind_request", version = "1.1.0", optional = true }
|
||||||
walkdir = "2"
|
walkdir = "2"
|
||||||
webauthn-rs = "0.2.5"
|
|
||||||
xdg = "2.2"
|
xdg = "2.2"
|
||||||
zstd = { version = "0.4", features = [ "bindgen" ] }
|
|
||||||
nom = "5.1"
|
nom = "5.1"
|
||||||
crossbeam-channel = "0.5"
|
crossbeam-channel = "0.5"
|
||||||
|
|
||||||
|
# Used only by examples currently:
|
||||||
|
zstd = { version = "0.6", features = [ "bindgen" ] }
|
||||||
|
|
||||||
|
pathpatterns = "0.1.2"
|
||||||
|
pxar = { version = "0.10.1", features = [ "tokio-io" ] }
|
||||||
|
|
||||||
|
proxmox = { version = "0.15.3", features = [ "sortable-macro" ] }
|
||||||
|
proxmox-http = { version = "0.5.4", features = [ "client", "http-helpers", "websocket" ] }
|
||||||
|
proxmox-io = "1"
|
||||||
|
proxmox-lang = "1"
|
||||||
|
proxmox-router = { version = "1.1", features = [ "cli" ] }
|
||||||
|
proxmox-schema = { version = "1", features = [ "api-macro" ] }
|
||||||
|
proxmox-section-config = "1"
|
||||||
|
proxmox-tfa = { version = "1.3", features = [ "api", "api-types" ] }
|
||||||
|
proxmox-time = "1"
|
||||||
|
proxmox-uuid = "1"
|
||||||
|
proxmox-shared-memory = "0.1.1"
|
||||||
|
proxmox-sys = "0.1.2"
|
||||||
|
|
||||||
|
proxmox-acme-rs = "0.3"
|
||||||
|
proxmox-apt = "0.8.0"
|
||||||
|
proxmox-async = "0.2"
|
||||||
|
proxmox-openid = "0.9.0"
|
||||||
|
|
||||||
|
pbs-api-types = { path = "pbs-api-types" }
|
||||||
|
pbs-buildcfg = { path = "pbs-buildcfg" }
|
||||||
|
pbs-client = { path = "pbs-client" }
|
||||||
|
pbs-config = { path = "pbs-config" }
|
||||||
|
pbs-datastore = { path = "pbs-datastore" }
|
||||||
|
proxmox-rest-server = { path = "proxmox-rest-server" }
|
||||||
|
proxmox-rrd = { path = "proxmox-rrd" }
|
||||||
|
pbs-tools = { path = "pbs-tools" }
|
||||||
|
pbs-tape = { path = "pbs-tape" }
|
||||||
|
|
||||||
|
# Local path overrides
|
||||||
|
# NOTE: You must run `cargo update` after changing this for it to take effect!
|
||||||
|
[patch.crates-io]
|
||||||
|
#proxmox = { path = "../proxmox/proxmox" }
|
||||||
|
#proxmox-http = { path = "../proxmox/proxmox-http" }
|
||||||
|
#proxmox-tfa = { path = "../proxmox/proxmox-tfa" }
|
||||||
|
#proxmox-schema = { path = "../proxmox/proxmox-schema" }
|
||||||
|
#pxar = { path = "../pxar" }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
default = []
|
default = []
|
||||||
#valgrind = ["valgrind_request"]
|
#valgrind = ["valgrind_request"]
|
||||||
|
112
Makefile
112
Makefile
@ -17,7 +17,8 @@ USR_BIN := \
|
|||||||
|
|
||||||
# Binaries usable by admins
|
# Binaries usable by admins
|
||||||
USR_SBIN := \
|
USR_SBIN := \
|
||||||
proxmox-backup-manager
|
proxmox-backup-manager \
|
||||||
|
proxmox-backup-debug \
|
||||||
|
|
||||||
# Binaries for services:
|
# Binaries for services:
|
||||||
SERVICE_BIN := \
|
SERVICE_BIN := \
|
||||||
@ -30,6 +31,23 @@ SERVICE_BIN := \
|
|||||||
RESTORE_BIN := \
|
RESTORE_BIN := \
|
||||||
proxmox-restore-daemon
|
proxmox-restore-daemon
|
||||||
|
|
||||||
|
SUBCRATES := \
|
||||||
|
pbs-api-types \
|
||||||
|
pbs-buildcfg \
|
||||||
|
pbs-client \
|
||||||
|
pbs-config \
|
||||||
|
pbs-datastore \
|
||||||
|
pbs-fuse-loop \
|
||||||
|
proxmox-rest-server \
|
||||||
|
proxmox-rrd \
|
||||||
|
pbs-tape \
|
||||||
|
pbs-tools \
|
||||||
|
proxmox-backup-banner \
|
||||||
|
proxmox-backup-client \
|
||||||
|
proxmox-file-restore \
|
||||||
|
proxmox-restore-daemon \
|
||||||
|
pxar-bin
|
||||||
|
|
||||||
ifeq ($(BUILD_MODE), release)
|
ifeq ($(BUILD_MODE), release)
|
||||||
CARGO_BUILD_ARGS += --release
|
CARGO_BUILD_ARGS += --release
|
||||||
COMPILEDIR := target/release
|
COMPILEDIR := target/release
|
||||||
@ -57,13 +75,15 @@ RESTORE_DBG_DEB=proxmox-backup-file-restore-dbgsym_${DEB_VERSION}_${ARCH}.deb
|
|||||||
DOC_DEB=${PACKAGE}-docs_${DEB_VERSION}_all.deb
|
DOC_DEB=${PACKAGE}-docs_${DEB_VERSION}_all.deb
|
||||||
|
|
||||||
DEBS=${SERVER_DEB} ${SERVER_DBG_DEB} ${CLIENT_DEB} ${CLIENT_DBG_DEB} \
|
DEBS=${SERVER_DEB} ${SERVER_DBG_DEB} ${CLIENT_DEB} ${CLIENT_DBG_DEB} \
|
||||||
${RESTORE_DEB} ${RESTORE_DBG_DEB}
|
${RESTORE_DEB} ${RESTORE_DBG_DEB} ${DEBUG_DEB} ${DEBUG_DBG_DEB}
|
||||||
|
|
||||||
DSC = rust-${PACKAGE}_${DEB_VERSION}.dsc
|
DSC = rust-${PACKAGE}_${DEB_VERSION}.dsc
|
||||||
|
|
||||||
DESTDIR=
|
DESTDIR=
|
||||||
|
|
||||||
all: cargo-build $(SUBDIRS)
|
tests ?= --workspace
|
||||||
|
|
||||||
|
all: $(SUBDIRS)
|
||||||
|
|
||||||
.PHONY: $(SUBDIRS)
|
.PHONY: $(SUBDIRS)
|
||||||
$(SUBDIRS):
|
$(SUBDIRS):
|
||||||
@ -75,19 +95,23 @@ test:
|
|||||||
$(CARGO) test $(tests) $(CARGO_BUILD_ARGS)
|
$(CARGO) test $(tests) $(CARGO_BUILD_ARGS)
|
||||||
|
|
||||||
doc:
|
doc:
|
||||||
$(CARGO) doc --no-deps $(CARGO_BUILD_ARGS)
|
$(CARGO) doc --workspace --no-deps $(CARGO_BUILD_ARGS)
|
||||||
|
|
||||||
# always re-create this dir
|
# always re-create this dir
|
||||||
.PHONY: build
|
.PHONY: build
|
||||||
build:
|
build:
|
||||||
|
@echo "Setting pkg-buildcfg version to: $(DEB_VERSION_UPSTREAM)"
|
||||||
|
sed -i -e 's/^version =.*$$/version = "$(DEB_VERSION_UPSTREAM)"/' \
|
||||||
|
pbs-buildcfg/Cargo.toml
|
||||||
rm -rf build
|
rm -rf build
|
||||||
rm -f debian/control
|
mkdir build
|
||||||
debcargo package --config debian/debcargo.toml --changelog-ready --no-overlay-write-back --directory build proxmox-backup $(shell dpkg-parsechangelog -l debian/changelog -SVersion | sed -e 's/-.*//')
|
cp -a debian \
|
||||||
sed -e '1,/^$$/ ! d' build/debian/control > build/debian/control.src
|
Cargo.toml src \
|
||||||
cat build/debian/control.src build/debian/control.in > build/debian/control
|
$(SUBCRATES) \
|
||||||
rm build/debian/control.in build/debian/control.src
|
docs etc examples tests www zsh-completions \
|
||||||
cp build/debian/control debian/control
|
defines.mk Makefile \
|
||||||
rm build/Cargo.lock
|
./build/
|
||||||
|
rm -f build/Cargo.lock
|
||||||
find build/debian -name "*.hint" -delete
|
find build/debian -name "*.hint" -delete
|
||||||
$(foreach i,$(SUBDIRS), \
|
$(foreach i,$(SUBDIRS), \
|
||||||
$(MAKE) -C build/$(i) clean ;)
|
$(MAKE) -C build/$(i) clean ;)
|
||||||
@ -107,7 +131,9 @@ deb: build
|
|||||||
lintian $(DEBS)
|
lintian $(DEBS)
|
||||||
|
|
||||||
.PHONY: deb-all
|
.PHONY: deb-all
|
||||||
deb-all: $(DOC_DEB) $(DEBS)
|
deb-all: build
|
||||||
|
cd build; dpkg-buildpackage -b -us -uc --no-pre-clean
|
||||||
|
lintian $(DEBS) $(DOC_DEB)
|
||||||
|
|
||||||
.PHONY: dsc
|
.PHONY: dsc
|
||||||
dsc: $(DSC)
|
dsc: $(DSC)
|
||||||
@ -115,27 +141,61 @@ $(DSC): build
|
|||||||
cd build; dpkg-buildpackage -S -us -uc -d -nc
|
cd build; dpkg-buildpackage -S -us -uc -d -nc
|
||||||
lintian $(DSC)
|
lintian $(DSC)
|
||||||
|
|
||||||
|
.PHONY: clean distclean deb clean
|
||||||
distclean: clean
|
distclean: clean
|
||||||
|
clean: clean-deb
|
||||||
clean:
|
|
||||||
$(foreach i,$(SUBDIRS), \
|
$(foreach i,$(SUBDIRS), \
|
||||||
$(MAKE) -C $(i) clean ;)
|
$(MAKE) -C $(i) clean ;)
|
||||||
$(CARGO) clean
|
$(CARGO) clean
|
||||||
rm -rf *.deb *.dsc *.tar.gz *.buildinfo *.changes build
|
rm -f .do-cargo-build
|
||||||
find . -name '*~' -exec rm {} ';'
|
find . -name '*~' -exec rm {} ';'
|
||||||
|
|
||||||
|
# allows one to avoid running cargo clean when one just wants to tidy up after a packgae build
|
||||||
|
clean-deb:
|
||||||
|
rm -rf *.deb *.dsc *.tar.gz *.buildinfo *.changes build/
|
||||||
|
|
||||||
.PHONY: dinstall
|
.PHONY: dinstall
|
||||||
dinstall: ${SERVER_DEB} ${SERVER_DBG_DEB} ${CLIENT_DEB} ${CLIENT_DBG_DEB}
|
dinstall: ${SERVER_DEB} ${SERVER_DBG_DEB} ${CLIENT_DEB} ${CLIENT_DBG_DEB} \
|
||||||
|
${DEBUG_DEB} ${DEBUG_DBG_DEB}
|
||||||
dpkg -i $^
|
dpkg -i $^
|
||||||
|
|
||||||
# make sure we build binaries before docs
|
# make sure we build binaries before docs
|
||||||
docs: cargo-build
|
docs: $(COMPILEDIR)/dump-catalog-shell-cli $(COMPILEDIR)/docgen
|
||||||
|
|
||||||
.PHONY: cargo-build
|
.PHONY: cargo-build
|
||||||
cargo-build:
|
cargo-build:
|
||||||
$(CARGO) build $(CARGO_BUILD_ARGS)
|
rm -f .do-cargo-build
|
||||||
|
$(MAKE) $(COMPILED_BINS)
|
||||||
|
|
||||||
|
$(COMPILED_BINS) $(COMPILEDIR)/dump-catalog-shell-cli $(COMPILEDIR)/docgen: .do-cargo-build
|
||||||
|
.do-cargo-build:
|
||||||
|
$(CARGO) build $(CARGO_BUILD_ARGS) \
|
||||||
|
--package proxmox-backup-banner \
|
||||||
|
--bin proxmox-backup-banner \
|
||||||
|
--package proxmox-backup-client \
|
||||||
|
--bin proxmox-backup-client \
|
||||||
|
--bin dump-catalog-shell-cli \
|
||||||
|
--bin proxmox-backup-debug \
|
||||||
|
--package proxmox-file-restore \
|
||||||
|
--bin proxmox-file-restore \
|
||||||
|
--package pxar-bin \
|
||||||
|
--bin pxar \
|
||||||
|
--package pbs-tape \
|
||||||
|
--bin pmt \
|
||||||
|
--bin pmtx \
|
||||||
|
--package proxmox-restore-daemon \
|
||||||
|
--bin proxmox-restore-daemon \
|
||||||
|
--package proxmox-backup \
|
||||||
|
--bin docgen \
|
||||||
|
--bin proxmox-backup-api \
|
||||||
|
--bin proxmox-backup-manager \
|
||||||
|
--bin proxmox-backup-proxy \
|
||||||
|
--bin proxmox-daily-update \
|
||||||
|
--bin proxmox-file-restore \
|
||||||
|
--bin proxmox-tape \
|
||||||
|
--bin sg-tape-cmd
|
||||||
|
touch "$@"
|
||||||
|
|
||||||
$(COMPILED_BINS): cargo-build
|
|
||||||
|
|
||||||
.PHONY: lint
|
.PHONY: lint
|
||||||
lint:
|
lint:
|
||||||
@ -161,12 +221,16 @@ install: $(COMPILED_BINS)
|
|||||||
install -m755 $(COMPILEDIR)/$(i) $(DESTDIR)$(LIBEXECDIR)/proxmox-backup/ ;)
|
install -m755 $(COMPILEDIR)/$(i) $(DESTDIR)$(LIBEXECDIR)/proxmox-backup/ ;)
|
||||||
$(MAKE) -C www install
|
$(MAKE) -C www install
|
||||||
$(MAKE) -C docs install
|
$(MAKE) -C docs install
|
||||||
|
ifeq (,$(filter nocheck,$(DEB_BUILD_OPTIONS)))
|
||||||
|
$(MAKE) test # HACK, only test now to avoid clobbering build files with wrong config
|
||||||
|
endif
|
||||||
|
|
||||||
.PHONY: upload
|
.PHONY: upload
|
||||||
upload: ${SERVER_DEB} ${CLIENT_DEB} ${RESTORE_DEB} ${DOC_DEB}
|
upload: ${SERVER_DEB} ${CLIENT_DEB} ${RESTORE_DEB} ${DOC_DEB} ${DEBUG_DEB}
|
||||||
# check if working directory is clean
|
# check if working directory is clean
|
||||||
git diff --exit-code --stat && git diff --exit-code --stat --staged
|
git diff --exit-code --stat && git diff --exit-code --stat --staged
|
||||||
tar cf - ${SERVER_DEB} ${SERVER_DBG_DEB} ${DOC_DEB} ${CLIENT_DEB} ${CLIENT_DBG_DEB} | \
|
tar cf - ${SERVER_DEB} ${SERVER_DBG_DEB} ${DOC_DEB} ${CLIENT_DEB} \
|
||||||
ssh -X repoman@repo.proxmox.com upload --product pbs --dist buster
|
${CLIENT_DBG_DEB} ${DEBUG_DEB} ${DEBUG_DBG_DEB} \
|
||||||
tar cf - ${CLIENT_DEB} ${CLIENT_DBG_DEB} | ssh -X repoman@repo.proxmox.com upload --product "pve,pmg" --dist buster
|
| ssh -X repoman@repo.proxmox.com upload --product pbs --dist bullseye
|
||||||
tar cf - ${RESTORE_DEB} ${RESTORE_DBG_DEB} | ssh -X repoman@repo.proxmox.com upload --product "pve" --dist buster
|
tar cf - ${CLIENT_DEB} ${CLIENT_DBG_DEB} | ssh -X repoman@repo.proxmox.com upload --product "pve,pmg,pbs-client" --dist bullseye
|
||||||
|
tar cf - ${RESTORE_DEB} ${RESTORE_DBG_DEB} | ssh -X repoman@repo.proxmox.com upload --product "pve" --dist bullseye
|
||||||
|
587
debian/changelog
vendored
587
debian/changelog
vendored
@ -1,3 +1,590 @@
|
|||||||
|
rust-proxmox-backup (2.1.2-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* docs: backup-client: fix wrong reference
|
||||||
|
|
||||||
|
* docs: remotes: note that protected flags will not be synced
|
||||||
|
|
||||||
|
* sync job: correctly apply rate limit
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 23 Nov 2021 13:56:15 +0100
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.1.1-2) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* docs: update and add traffic control related screenshots
|
||||||
|
|
||||||
|
* docs: mention traffic control (bandwidth limits) for sync jobs
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Mon, 22 Nov 2021 16:07:39 +0100
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.1.1-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* fix proxmox-backup-manager sync-job list
|
||||||
|
|
||||||
|
* ui, api: sync-job: allow one to configure a rate limit
|
||||||
|
|
||||||
|
* api: snapshot list: set default for 'protected' flag
|
||||||
|
|
||||||
|
* ui: datastore content: rework rendering protection state
|
||||||
|
|
||||||
|
* docs: update traffic control docs (use HumanBytes)
|
||||||
|
|
||||||
|
* ui: traffic-control: include ipv6 in 'all' networks
|
||||||
|
|
||||||
|
* ui: traffic-control edit: add spaces between networks for more
|
||||||
|
readabillity
|
||||||
|
|
||||||
|
* tape: fix passing-through key-fingerprint
|
||||||
|
|
||||||
|
* avoid a bogus error regarding logrotate-path due to a reversed check
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Mon, 22 Nov 2021 12:24:31 +0100
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.1.0-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* rest server: make successful-ticket auth log a debug one to avoid
|
||||||
|
syslog spam
|
||||||
|
|
||||||
|
* traffic-controls: add API/CLI to show current traffic
|
||||||
|
|
||||||
|
* docs: add traffic control section
|
||||||
|
|
||||||
|
* ui: use TFA widgets from widget toolkit
|
||||||
|
|
||||||
|
* sync: allow pulling groups selectively
|
||||||
|
|
||||||
|
* fix #3533: tape backup: filter groups according to config
|
||||||
|
|
||||||
|
* proxmox-tape: add missing notify-user option to backup command
|
||||||
|
|
||||||
|
* openid: allow arbitrary username-claims
|
||||||
|
|
||||||
|
* openid: support configuring the prompt, scopes and ACR values
|
||||||
|
|
||||||
|
* use human-byte for traffic-control rate-in/out and burst-in/out config
|
||||||
|
|
||||||
|
* ui: add traffic control view and editor
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Sat, 20 Nov 2021 22:44:07 +0100
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.14-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* fix directory permission problems
|
||||||
|
|
||||||
|
* add traffic control configuration config with API
|
||||||
|
|
||||||
|
* proxmox-backup-proxy: implement traffic control
|
||||||
|
|
||||||
|
* proxmox-backup-client: add rate/burst parameter to backup/restore CLI
|
||||||
|
|
||||||
|
* openid_login: vertify that firstname, lastname and email fits our
|
||||||
|
schema definitions
|
||||||
|
|
||||||
|
* docs: add info about protection flag to client docs
|
||||||
|
|
||||||
|
* fix #3602: ui: datastore/Content: add action to set protection status
|
||||||
|
|
||||||
|
* ui: add protected icon to snapshot (if they are protected)
|
||||||
|
|
||||||
|
* ui: PruneInputPanel: add keepReason 'protected' for protected backups
|
||||||
|
|
||||||
|
* proxmox-backup-client: add 'protected' commands
|
||||||
|
|
||||||
|
* acme: interpret no TOS as accepted
|
||||||
|
|
||||||
|
* acme: new_account: prevent replacing existing accounts
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Fri, 12 Nov 2021 08:04:55 +0100
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.13-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* tape: simplify export_media_set for pool writer
|
||||||
|
|
||||||
|
* tape: improve export_media error message for not found tape
|
||||||
|
|
||||||
|
* rest-server: use hashmap for parameter errors
|
||||||
|
|
||||||
|
* proxmox-rrd: use new file firmat with higher resolution
|
||||||
|
|
||||||
|
* proxmox-rrd: use a journal to reduce amount of bytes written
|
||||||
|
|
||||||
|
* use new fsync parameter to replace_file and atomic_open_or_create
|
||||||
|
|
||||||
|
* docs: langauge and formatting fixup
|
||||||
|
|
||||||
|
* docs: Update for new features/functionality
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Thu, 21 Oct 2021 08:17:00 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.12-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* proxmox-backup-proxy: clean up old tasks when their reference was rotated
|
||||||
|
out of the task-log index
|
||||||
|
|
||||||
|
* api daemons: fix sending log-reopen command
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 19 Oct 2021 10:48:28 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.11-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* drop aritifical limits for task-UPID length
|
||||||
|
|
||||||
|
* tools: smart: only throw error for the fatal usage errors of smartctl
|
||||||
|
|
||||||
|
* api: improve returning errors for extjs formatter
|
||||||
|
|
||||||
|
* proxmox-rest-server: improve logging
|
||||||
|
|
||||||
|
* subscription: switch verification domain over to shop.proxmox.com
|
||||||
|
|
||||||
|
* rest-server/daemon: use new sd_notify_barrier helper for handling
|
||||||
|
synchronization with systemd on service reloading
|
||||||
|
|
||||||
|
* ui: datastore/Content: add empty text for no snapshots
|
||||||
|
|
||||||
|
* ui: datastore/Content: move first store-load into activate listener to
|
||||||
|
ensure we've a proper loading mask for better UX
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 05 Oct 2021 16:34:14 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.10-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* ui: fix order of prune keep reasons
|
||||||
|
|
||||||
|
* server: add proxmox-backup-debug binary with chunk/file inspection, an API
|
||||||
|
shell with completion support
|
||||||
|
|
||||||
|
* restructured code base to reduce linkage and libraray ABI version
|
||||||
|
constraints for all non-server binaries (client, pxar, file-restore)
|
||||||
|
|
||||||
|
* zsh: fix passign parameters in auto-completion scripts
|
||||||
|
|
||||||
|
* tape: also add 'force-media-set' to availablea CLI options
|
||||||
|
|
||||||
|
* api: nodes: add missing node list (index) api endpoint
|
||||||
|
|
||||||
|
* docs: proxmox-backup-debug: add info about the new 'api' subcommand
|
||||||
|
|
||||||
|
* docs/technical-overview: add troubleshooting section
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 21 Sep 2021 14:00:48 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.9-2) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* tape backup: mention groups that were empty
|
||||||
|
|
||||||
|
* tape: compute next-media-label for each tape backup job
|
||||||
|
|
||||||
|
* tape: lto: increase default timeout to 10 minutes
|
||||||
|
|
||||||
|
* ui: display next-media-label for tape backup jobs
|
||||||
|
|
||||||
|
* cli: proxmox-tape backup-job list: use status api and display next-run
|
||||||
|
and next-media-label
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 24 Aug 2021 14:44:12 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.8-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* use proxmox-apt to 0.6
|
||||||
|
|
||||||
|
* api: apt: adapt to proxmox-apt back-end changes
|
||||||
|
|
||||||
|
* api/ui: allow zstd compression for new zpools
|
||||||
|
|
||||||
|
* tape: media_catalog: add snapshot list cache for catalog
|
||||||
|
|
||||||
|
* api2: tape: media: use MediaCatalog::snapshot_list for content listing
|
||||||
|
|
||||||
|
* tape: lock media_catalog file to to get a consistent view with load_catalog
|
||||||
|
|
||||||
|
* tape: changer: handle libraries that sends wrong amount of data
|
||||||
|
|
||||||
|
* tape: changer: remove unnecesary inquiry parameter
|
||||||
|
|
||||||
|
* api2: tape/restore: commit temporary catalog at the end
|
||||||
|
|
||||||
|
* docs: tape: add instructions on how to restore the catalog
|
||||||
|
|
||||||
|
* ui: tape/ChangerStatus: improve layout for large libraries
|
||||||
|
|
||||||
|
* tape: changer: handle invalid descriptor data from library in status page
|
||||||
|
|
||||||
|
* datastore config: cleanup code (use flatten attribute)
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Mon, 02 Aug 2021 10:34:55 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.7-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* tape changer: better cope with models that are not following spec
|
||||||
|
proposals when returning the status page
|
||||||
|
|
||||||
|
* tape changer: make DVCID information optional, not all devices return it
|
||||||
|
|
||||||
|
* restore daemon: setup the 'backup' system user and group in the minimal
|
||||||
|
restore environment, as we like to ensure that all state files are ownend
|
||||||
|
by them.
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Fri, 23 Jul 2021 08:43:51 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.6-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* increase maximum drives per changer to 255
|
||||||
|
|
||||||
|
* allow one to pass a secret not only directly through the environment value,
|
||||||
|
but also indirectly through a file path, an open file descriptor or a
|
||||||
|
command that can write the secret to standard out.
|
||||||
|
|
||||||
|
* pull in new proxmox library version to improve the file system
|
||||||
|
comaptibility on creation of atomic files, e.g., lock files.
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Thu, 22 Jul 2021 10:22:19 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.5-2) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* ui: tape: backup overview: increase timeout for media-set content
|
||||||
|
|
||||||
|
* tape: changer: always retry until timeout
|
||||||
|
|
||||||
|
* file-restore: increase lock timeout on QEMU map
|
||||||
|
|
||||||
|
* fix #3515: file-restore-daemon: allow LVs/PVs with dash in name
|
||||||
|
|
||||||
|
* fix #3526: correctly filter tasks with 'since' and 'until'
|
||||||
|
|
||||||
|
* tape: changer: make scsi request for DVCID a separate one, as some
|
||||||
|
libraries cannot handle requesting that combined with volume tags in one
|
||||||
|
go
|
||||||
|
|
||||||
|
* api, ui: datastore: add new 'prune-datastore' api call and expose it with
|
||||||
|
a 'Prune All' button
|
||||||
|
|
||||||
|
* make creating log files more robust so that theys are always owned by the
|
||||||
|
less privileged `backup` user
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Wed, 21 Jul 2021 09:12:39 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.4-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* change tape drive lock path to avoid issues with sticky bit on tmpfs
|
||||||
|
mountpoint
|
||||||
|
|
||||||
|
* tape: changer: query transport-element types separately
|
||||||
|
|
||||||
|
* auth: improve thread safety of 'crypt' C-library
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Mon, 12 Jul 2021 18:51:21 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.3-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* api: apt: add repositories info and update calls
|
||||||
|
|
||||||
|
* ui: administration: add APT repositories status and update panel
|
||||||
|
|
||||||
|
* api: access domains: add get/create/update/delete endpoints for realms
|
||||||
|
|
||||||
|
* ui: access control: add 'Realm' tab for adding and editing OpenID Connect
|
||||||
|
identity provider
|
||||||
|
|
||||||
|
* fix #3447: ui: Dashboard: disallow selection of datastore statistics row
|
||||||
|
|
||||||
|
* ui: tapeRestore: make window non-resizable
|
||||||
|
|
||||||
|
* ui: dashboard: rework resource-load panel to a more detailed status panel,
|
||||||
|
showing, among other things, uptime, Kernel version, CPU info and
|
||||||
|
repository status.
|
||||||
|
|
||||||
|
* ui: adminsitration/dashboard: auto-scale columns count and add
|
||||||
|
browser-local setting to override that to a fixed value of columns.
|
||||||
|
|
||||||
|
* fix #3212: api, ui: add support for notes on backup groups
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Mon, 12 Jul 2021 08:07:41 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.2-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* ui: use task list component from widget toolkit
|
||||||
|
|
||||||
|
* api: add keep-job-configs flag to datastore remove endpoint
|
||||||
|
|
||||||
|
* api: config: delete datastore: also remove tape backup jobs
|
||||||
|
|
||||||
|
* ui: tape restore: mark datastore selector as 'not a form field' to fix
|
||||||
|
compatibility with ExtJS 7.0
|
||||||
|
|
||||||
|
* ui: datastore removal: only navigate away when the user actually confirmed
|
||||||
|
the removal of that datastore
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Thu, 08 Jul 2021 14:44:12 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.1-2) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* file restore daemon: log basic startup steps
|
||||||
|
|
||||||
|
* REST-API: set error message extension for bad-request response log to
|
||||||
|
ensure the actual error is logged in any (access) log, making debugging
|
||||||
|
such issues easier
|
||||||
|
|
||||||
|
* restore daemon: create /run/proxmox-backup on startup as there's now some
|
||||||
|
runtime state saved there, which failed all API requests to the restore
|
||||||
|
daemon otherwise
|
||||||
|
|
||||||
|
* restore daemon: use millisecond log resolution
|
||||||
|
|
||||||
|
* fix #3496: acme: plugin: actually sleep after setting the TXT record,
|
||||||
|
ensuring DNS propagation of that record. This makes it catch up with the
|
||||||
|
docs/web-interface, where the option was already available.
|
||||||
|
|
||||||
|
* docs: initial update to repositories for bullseye
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Sat, 03 Jul 2021 23:14:49 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.0-2) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* file-restore-daemon/disk: add LVM (thin) support
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Sat, 03 Jul 2021 02:15:16 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (2.0.0-1) bullseye; urgency=medium
|
||||||
|
|
||||||
|
* initial bump for Debian 11 Bullseye / Proxmox Backup Server 2.0
|
||||||
|
|
||||||
|
* ui: datastore list summary: catch and show errors per datastore
|
||||||
|
|
||||||
|
* ui: dashboard: task summary: add a 'close' tool to the header
|
||||||
|
|
||||||
|
* ensure that backups which are currently being restored or backed up to a
|
||||||
|
tape won't get pruned
|
||||||
|
|
||||||
|
* improve error handling when locking a tape drive for a backup job
|
||||||
|
|
||||||
|
* client/pull: log snapshots that are skipped because of creation time being
|
||||||
|
older than last sync time
|
||||||
|
|
||||||
|
* ui: datastore options: add remove button to drop a datastore from the
|
||||||
|
configuration, without removing any actual data
|
||||||
|
|
||||||
|
* ui: tape: drive selector: do not auto select the drive
|
||||||
|
|
||||||
|
* ui: tape: backup job: use correct default value for pbsUserSelector
|
||||||
|
|
||||||
|
* fix #3433: disks: port over Proxmox VE's S.M.A.R.T wearout logic
|
||||||
|
|
||||||
|
* backup: add helpers for async last recently used (LRU) caches for chunk
|
||||||
|
and index reading of backup snapshot
|
||||||
|
|
||||||
|
* fix #3459: manager: add --ignore-verified and --outdated-after parameters
|
||||||
|
|
||||||
|
* proxmox-backup-manager: show task log on datastore create
|
||||||
|
|
||||||
|
* tape: snapshot reader: read chunks sorted by inode (per index) to improve
|
||||||
|
sequential reads when backing up data from slow spinning disks to tape.
|
||||||
|
|
||||||
|
* file-restore: support ZFS pools
|
||||||
|
|
||||||
|
* improve fix for #3393: pxar create: try to read xattrs/fcaps/acls by default
|
||||||
|
|
||||||
|
* fix compatibility with ExtJS 7.0
|
||||||
|
|
||||||
|
* docs: build api-viewer from widget-toolkit-dev
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Mon, 28 Jun 2021 19:35:40 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.9-1) stable; urgency=medium
|
||||||
|
|
||||||
|
* lto/sg_tape/encryption: remove non lto-4 supported byte
|
||||||
|
|
||||||
|
* ui: improve tape restore
|
||||||
|
|
||||||
|
* ui: panel/UsageChart: change downloadServerUrl
|
||||||
|
|
||||||
|
* ui: css fixes and cleanups
|
||||||
|
|
||||||
|
* api2/tape: add api call to list media sets
|
||||||
|
|
||||||
|
* ui: tape/BackupOverview: expand pools by default
|
||||||
|
|
||||||
|
* api: node/journal: fix parameter extraction of /nodes/node/journal
|
||||||
|
|
||||||
|
* file-restore-daemon: limit concurrent download calls
|
||||||
|
|
||||||
|
* file-restore-daemon: watchdog: add inhibit for long downloads
|
||||||
|
|
||||||
|
* file-restore-daemon: work around tokio DuplexStream bug
|
||||||
|
|
||||||
|
* apt: fix removal of non-existant http-proxy config
|
||||||
|
|
||||||
|
* file-restore-daemon: disk: add RawFs bucket type
|
||||||
|
|
||||||
|
* file-restore-daemon: disk: ignore "invalid fs" error
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 01 Jun 2021 08:24:01 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.8-1) stable; urgency=medium
|
||||||
|
|
||||||
|
* api-proxy: implement 'reload-certificate' command and hot-reload proxy
|
||||||
|
certificate when updating via the API
|
||||||
|
|
||||||
|
* ui: add task descriptions for ACME/Let's Encrypt related tasks
|
||||||
|
|
||||||
|
* correctly set apt proxy configuration
|
||||||
|
|
||||||
|
* ui: configuration: support setting a HTTP proxy for APT and subscription
|
||||||
|
checks.
|
||||||
|
|
||||||
|
* ui: tape: add 'Force new Media-Set' checkbox to manual backup
|
||||||
|
|
||||||
|
* ui: datastore/Content: add forget (delete) button for whole backup groups
|
||||||
|
|
||||||
|
* ui: tape: backup overview: move restore buttons inline to action-buttons,
|
||||||
|
making the UX more similar to the datastore content tree-view
|
||||||
|
|
||||||
|
* ui: tape restore: enabling selecting multiple snapshots
|
||||||
|
|
||||||
|
* ui: dashboards statistics: visualize datastores where querying the usage
|
||||||
|
failed
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Fri, 21 May 2021 18:21:28 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.7-1) unstable; urgency=medium
|
||||||
|
|
||||||
|
* client: use stderr for all fingerprint confirm msgs
|
||||||
|
|
||||||
|
* fix #3391: improve mismatched fingerprint handling
|
||||||
|
|
||||||
|
* tape: add single snapshot restore
|
||||||
|
|
||||||
|
* docs/api-viewer: improve rendering of array format
|
||||||
|
|
||||||
|
* tape/pool_writer: do not unwrap on channel send
|
||||||
|
|
||||||
|
* ui: window/SyncJobEdit: disable autoSelect for remote datastore
|
||||||
|
|
||||||
|
* ui: tape: rename 'Datastore' to 'Target Datastore'
|
||||||
|
|
||||||
|
* manager: acme plugin: auto-complete available DNS challenge types
|
||||||
|
|
||||||
|
* manager: acme plugin: remove ID completion helper from add command
|
||||||
|
|
||||||
|
* completion: ACME plugin type: comment out http type for now, not useful
|
||||||
|
|
||||||
|
* acme: use proxmox-acme-plugins and load schema from there
|
||||||
|
|
||||||
|
* fix 3296: add http_proxy to node config, and provide a cli
|
||||||
|
|
||||||
|
* fix #3331: improve progress for last snapshot in group
|
||||||
|
|
||||||
|
* file-restore: add debug mode with serial access
|
||||||
|
|
||||||
|
* file-restore: support more drives
|
||||||
|
|
||||||
|
* file-restore: add more RAM for VMs with many drives or debug
|
||||||
|
|
||||||
|
* file-restore: try to kill VM when stale
|
||||||
|
|
||||||
|
* make sure URI paths start with a slash
|
||||||
|
|
||||||
|
* tape: use LOCATE(16) SCSI command
|
||||||
|
|
||||||
|
* call create_run_dir() at daemon startup
|
||||||
|
|
||||||
|
* tape/drive: add 'move_to_file' to TapeDriver trait
|
||||||
|
|
||||||
|
* proxmox_restore_daemon: mount ntfs with 'utf8' option
|
||||||
|
|
||||||
|
* client/http_client: add necessary brackets for ipv6
|
||||||
|
|
||||||
|
* docs: tape: clarify LTO-4/5 support
|
||||||
|
|
||||||
|
* tape/restore: optimize chunk restore behaviour
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 11 May 2021 13:22:49 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.6-2) unstable; urgency=medium
|
||||||
|
|
||||||
|
* fix permissions set in create_run_dir
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 04 May 2021 12:25:00 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.6-1) unstable; urgency=medium
|
||||||
|
|
||||||
|
* tape restore: do not verify restored files
|
||||||
|
|
||||||
|
* tape restore: add restore speed to logs
|
||||||
|
|
||||||
|
* tape restore: write datastore in separate thread
|
||||||
|
|
||||||
|
* add ACME support
|
||||||
|
|
||||||
|
* add node config
|
||||||
|
|
||||||
|
* docs: user-management: add note about untrusted certificates for
|
||||||
|
webauthn
|
||||||
|
|
||||||
|
* bin: use extract_output_format where necessary
|
||||||
|
|
||||||
|
* add ctime and size function to IndexFile trait
|
||||||
|
|
||||||
|
* ui: tape: handle tapes in changers without barcode
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 04 May 2021 12:09:25 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.5-3) stable; urgency=medium
|
||||||
|
|
||||||
|
* file-restore: use 'norecovery' for XFS filesystem to allow mounting
|
||||||
|
those which where not un-mounted during backup
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Thu, 29 Apr 2021 15:26:13 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.5-2) stable; urgency=medium
|
||||||
|
|
||||||
|
* file-restore: strip .img.fidx suffix from drive serials to avoid running
|
||||||
|
in the 20 character limit SCSI serial values have.
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Wed, 28 Apr 2021 11:15:08 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.5-1) unstable; urgency=medium
|
||||||
|
|
||||||
|
* tools/sgutils2: add size workaround for mode_sense
|
||||||
|
|
||||||
|
* tape: add read_medium_configuration_page() to detect WORM media
|
||||||
|
|
||||||
|
* file-restore: fix package name for kernel/initramfs image
|
||||||
|
|
||||||
|
* tape: remove MediumType struct, which is only valid on IBM drives
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 27 Apr 2021 12:20:04 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.4-1) unstable; urgency=medium
|
||||||
|
|
||||||
|
* file-restore: add size to image files and components
|
||||||
|
|
||||||
|
* file-restore: exit with code 1 in case streaming fails
|
||||||
|
|
||||||
|
* file-restore: use less memory for VM (now 128 MiB) and reboot on panic
|
||||||
|
|
||||||
|
* ui: tape: improve reload drive-status logic on user actions
|
||||||
|
|
||||||
|
* tape backup: list the snapshots we could back up on failed backup
|
||||||
|
notification
|
||||||
|
|
||||||
|
* Improve on a scheduling issue when updating the calendar event such, that
|
||||||
|
it would had triggered between the last-run and now. Use the next future
|
||||||
|
event as actual next trigger instead.
|
||||||
|
|
||||||
|
* SCSI mode sense: include the expected and unexpected sizes in the error
|
||||||
|
message, to allow easier debugging
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Tue, 27 Apr 2021 08:27:10 +0200
|
||||||
|
|
||||||
|
rust-proxmox-backup (1.1.3-2) unstable; urgency=medium
|
||||||
|
|
||||||
|
* improve check for LTO4 tapes
|
||||||
|
|
||||||
|
* api: node status: return further information about SWAP, IO-wait, CPU info
|
||||||
|
and Kernel version
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Fri, 23 Apr 2021 10:52:08 +0200
|
||||||
|
|
||||||
rust-proxmox-backup (1.1.3-1) unstable; urgency=medium
|
rust-proxmox-backup (1.1.3-1) unstable; urgency=medium
|
||||||
|
|
||||||
* tape restore: improve datastore locking when GC runs at the same time
|
* tape restore: improve datastore locking when GC runs at the same time
|
||||||
|
97
debian/control
vendored
97
debian/control
vendored
@ -1,29 +1,32 @@
|
|||||||
Source: rust-proxmox-backup
|
Source: rust-proxmox-backup
|
||||||
Section: admin
|
Section: admin
|
||||||
Priority: optional
|
Priority: optional
|
||||||
Build-Depends: debhelper (>= 11),
|
Build-Depends: debhelper (>= 12),
|
||||||
dh-cargo (>= 18),
|
dh-cargo (>= 24),
|
||||||
cargo:native,
|
cargo:native,
|
||||||
rustc:native,
|
rustc:native,
|
||||||
libstd-rust-dev,
|
libstd-rust-dev,
|
||||||
librust-anyhow-1+default-dev,
|
librust-anyhow-1+default-dev,
|
||||||
librust-apt-pkg-native-0.3+default-dev (>= 0.3.2-~~),
|
librust-apt-pkg-native-0.3+default-dev (>= 0.3.2-~~),
|
||||||
librust-base64-0.12+default-dev,
|
librust-base64-0.13+default-dev,
|
||||||
librust-bitflags-1+default-dev (>= 1.2.1-~~),
|
librust-bitflags-1+default-dev (>= 1.2.1-~~),
|
||||||
librust-bytes-1+default-dev,
|
librust-bytes-1+default-dev,
|
||||||
|
librust-cidr-0.2+default-dev (>= 0.2.1-~~),
|
||||||
librust-crc32fast-1+default-dev,
|
librust-crc32fast-1+default-dev,
|
||||||
librust-crossbeam-channel-0.5+default-dev,
|
librust-crossbeam-channel-0.5+default-dev,
|
||||||
librust-endian-trait-0.6+arrays-dev,
|
librust-endian-trait-0.6+arrays-dev,
|
||||||
librust-endian-trait-0.6+default-dev,
|
librust-endian-trait-0.6+default-dev,
|
||||||
librust-env-logger-0.7+default-dev,
|
librust-env-logger-0.7+default-dev,
|
||||||
librust-flate2-1+default-dev,
|
librust-flate2-1+default-dev,
|
||||||
|
librust-foreign-types-0.3+default-dev,
|
||||||
librust-futures-0.3+default-dev,
|
librust-futures-0.3+default-dev,
|
||||||
librust-h2-0.3+default-dev,
|
librust-h2-0.3+default-dev,
|
||||||
librust-h2-0.3+stream-dev,
|
librust-h2-0.3+stream-dev,
|
||||||
librust-handlebars-3+default-dev,
|
librust-handlebars-3+default-dev,
|
||||||
|
librust-hex-0.4+default-dev (>= 0.4.3-~~),
|
||||||
librust-http-0.2+default-dev,
|
librust-http-0.2+default-dev,
|
||||||
librust-hyper-0.14+default-dev,
|
librust-hyper-0.14+default-dev (>= 0.14.5-~~),
|
||||||
librust-hyper-0.14+full-dev,
|
librust-hyper-0.14+full-dev (>= 0.14.5-~~),
|
||||||
librust-lazy-static-1+default-dev (>= 1.4-~~),
|
librust-lazy-static-1+default-dev (>= 1.4-~~),
|
||||||
librust-libc-0.2+default-dev,
|
librust-libc-0.2+default-dev,
|
||||||
librust-log-0.4+default-dev,
|
librust-log-0.4+default-dev,
|
||||||
@ -36,51 +39,76 @@ Build-Depends: debhelper (>= 11),
|
|||||||
librust-pam-sys-0.5+default-dev,
|
librust-pam-sys-0.5+default-dev,
|
||||||
librust-pathpatterns-0.1+default-dev (>= 0.1.2-~~),
|
librust-pathpatterns-0.1+default-dev (>= 0.1.2-~~),
|
||||||
librust-percent-encoding-2+default-dev (>= 2.1-~~),
|
librust-percent-encoding-2+default-dev (>= 2.1-~~),
|
||||||
librust-pin-project-1+default-dev,
|
librust-pin-project-lite-0.2+default-dev,
|
||||||
librust-pin-utils-0.1+default-dev,
|
librust-proxmox-0.15+default-dev (>= 0.15.3-~~),
|
||||||
librust-proxmox-0.11+api-macro-dev (>= 0.11.1-~~),
|
librust-proxmox-0.15+sortable-macro-dev (>= 0.15.3-~~),
|
||||||
librust-proxmox-0.11+default-dev (>= 0.11.1-~~),
|
librust-proxmox-0.15+tokio-dev (>= 0.15.3-~~),
|
||||||
librust-proxmox-0.11+sortable-macro-dev (>= 0.11.1-~~),
|
librust-proxmox-acme-rs-0.3+default-dev,
|
||||||
librust-proxmox-0.11+websocket-dev (>= 0.11.1-~~),
|
librust-proxmox-apt-0.8+default-dev,
|
||||||
|
librust-proxmox-async-0.2+default-dev,
|
||||||
|
librust-proxmox-borrow-1+default-dev,
|
||||||
librust-proxmox-fuse-0.1+default-dev (>= 0.1.1-~~),
|
librust-proxmox-fuse-0.1+default-dev (>= 0.1.1-~~),
|
||||||
|
librust-proxmox-http-0.5+client-dev (>= 0.5.4-~~),
|
||||||
|
librust-proxmox-http-0.5+default-dev (>= 0.5.4-~~),
|
||||||
|
librust-proxmox-http-0.5+http-helpers-dev (>= 0.5.4-~~),
|
||||||
|
librust-proxmox-http-0.5+websocket-dev (>= 0.5.4-~~),
|
||||||
|
librust-proxmox-io-1+default-dev,
|
||||||
|
librust-proxmox-io-1+tokio-dev,
|
||||||
|
librust-proxmox-lang-1+default-dev,
|
||||||
|
librust-proxmox-openid-0.9+default-dev,
|
||||||
|
librust-proxmox-router-1+cli-dev (>= 1.1-~~),
|
||||||
|
librust-proxmox-router-1+default-dev (>= 1.1-~~),
|
||||||
|
librust-proxmox-schema-1+api-macro-dev (>= 1.0.1-~~),
|
||||||
|
librust-proxmox-schema-1+default-dev (>= 1.0.1-~~),
|
||||||
|
librust-proxmox-schema-1+upid-api-impl-dev (>= 1.0.1-~~),
|
||||||
|
librust-proxmox-section-config-1+default-dev,
|
||||||
|
librust-proxmox-shared-memory-0.1+default-dev (>= 0.1.1-~~),
|
||||||
|
librust-proxmox-sys-0.1+default-dev (>= 0.1.2-~~),
|
||||||
|
librust-proxmox-tfa-1+api-dev (>= 1.3-~~),
|
||||||
|
librust-proxmox-tfa-1+api-types-dev (>= 1.3-~~),
|
||||||
|
librust-proxmox-tfa-1+default-dev (>= 1.3-~~),
|
||||||
|
librust-proxmox-time-1+default-dev (>= 1.1-~~),
|
||||||
|
librust-proxmox-uuid-1+default-dev,
|
||||||
|
librust-proxmox-uuid-1+serde-dev,
|
||||||
librust-pxar-0.10+default-dev (>= 0.10.1-~~),
|
librust-pxar-0.10+default-dev (>= 0.10.1-~~),
|
||||||
librust-pxar-0.10+tokio-io-dev (>= 0.10.1-~~),
|
librust-pxar-0.10+tokio-io-dev (>= 0.10.1-~~),
|
||||||
librust-regex-1+default-dev (>= 1.2-~~),
|
librust-regex-1+default-dev (>= 1.2-~~),
|
||||||
librust-rustyline-7+default-dev,
|
librust-rustyline-7+default-dev,
|
||||||
librust-serde-1+default-dev,
|
librust-serde-1+default-dev,
|
||||||
librust-serde-1+derive-dev,
|
librust-serde-1+derive-dev,
|
||||||
|
librust-serde-cbor-0.11+default-dev (>= 0.11.1-~~),
|
||||||
librust-serde-json-1+default-dev,
|
librust-serde-json-1+default-dev,
|
||||||
librust-siphasher-0.3+default-dev,
|
librust-siphasher-0.3+default-dev,
|
||||||
librust-syslog-4+default-dev,
|
librust-syslog-4+default-dev,
|
||||||
librust-thiserror-1+default-dev,
|
librust-thiserror-1+default-dev,
|
||||||
librust-tokio-1+default-dev,
|
librust-tokio-1+default-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+fs-dev,
|
librust-tokio-1+fs-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+io-std-dev,
|
librust-tokio-1+io-std-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+io-util-dev,
|
librust-tokio-1+io-util-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+macros-dev,
|
librust-tokio-1+macros-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+net-dev,
|
librust-tokio-1+net-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+parking-lot-dev,
|
librust-tokio-1+parking-lot-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+process-dev,
|
librust-tokio-1+process-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+rt-dev,
|
librust-tokio-1+rt-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+rt-multi-thread-dev,
|
librust-tokio-1+rt-multi-thread-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+signal-dev,
|
librust-tokio-1+signal-dev (>= 1.6-~~),
|
||||||
librust-tokio-1+time-dev,
|
librust-tokio-1+sync-dev (>= 1.6-~~),
|
||||||
|
librust-tokio-1+time-dev (>= 1.6-~~),
|
||||||
librust-tokio-openssl-0.6+default-dev (>= 0.6.1-~~),
|
librust-tokio-openssl-0.6+default-dev (>= 0.6.1-~~),
|
||||||
librust-tokio-stream-0.1+default-dev,
|
librust-tokio-stream-0.1+default-dev,
|
||||||
librust-tokio-util-0.6+codec-dev,
|
librust-tokio-util-0.6+codec-dev,
|
||||||
librust-tokio-util-0.6+default-dev,
|
librust-tokio-util-0.6+default-dev,
|
||||||
librust-tokio-util-0.6+io-dev,
|
librust-tokio-util-0.6+io-dev,
|
||||||
librust-tower-service-0.3+default-dev,
|
librust-tower-service-0.3+default-dev,
|
||||||
librust-udev-0.4+default-dev | librust-udev-0.3+default-dev,
|
librust-udev-0.4+default-dev,
|
||||||
librust-url-2+default-dev (>= 2.1-~~),
|
librust-url-2+default-dev (>= 2.1-~~),
|
||||||
librust-walkdir-2+default-dev,
|
librust-walkdir-2+default-dev,
|
||||||
librust-webauthn-rs-0.2+default-dev (>= 0.2.5-~~),
|
|
||||||
librust-xdg-2+default-dev (>= 2.2-~~),
|
librust-xdg-2+default-dev (>= 2.2-~~),
|
||||||
librust-zstd-0.4+bindgen-dev,
|
librust-zstd-0.6+bindgen-dev,
|
||||||
librust-zstd-0.4+default-dev,
|
librust-zstd-0.6+default-dev,
|
||||||
libacl1-dev,
|
libacl1-dev,
|
||||||
libfuse3-dev,
|
libfuse3-dev,
|
||||||
libsystemd-dev,
|
libsystemd-dev (>= 246-~~),
|
||||||
uuid-dev,
|
uuid-dev,
|
||||||
libsgutils2-dev,
|
libsgutils2-dev,
|
||||||
bash-completion,
|
bash-completion,
|
||||||
@ -91,6 +119,7 @@ Build-Depends: debhelper (>= 11),
|
|||||||
graphviz <!nodoc>,
|
graphviz <!nodoc>,
|
||||||
latexmk <!nodoc>,
|
latexmk <!nodoc>,
|
||||||
patchelf,
|
patchelf,
|
||||||
|
proxmox-widget-toolkit-dev <!nodoc>,
|
||||||
pve-eslint (>= 7.18.0-1),
|
pve-eslint (>= 7.18.0-1),
|
||||||
python3-docutils,
|
python3-docutils,
|
||||||
python3-pygments,
|
python3-pygments,
|
||||||
@ -101,16 +130,18 @@ Build-Depends: debhelper (>= 11),
|
|||||||
texlive-xetex <!nodoc>,
|
texlive-xetex <!nodoc>,
|
||||||
xindy <!nodoc>
|
xindy <!nodoc>
|
||||||
Maintainer: Proxmox Support Team <support@proxmox.com>
|
Maintainer: Proxmox Support Team <support@proxmox.com>
|
||||||
Standards-Version: 4.4.1
|
Standards-Version: 4.5.1
|
||||||
Vcs-Git: git://git.proxmox.com/git/proxmox-backup.git
|
Vcs-Git: git://git.proxmox.com/git/proxmox-backup.git
|
||||||
Vcs-Browser: https://git.proxmox.com/?p=proxmox-backup.git;a=summary
|
Vcs-Browser: https://git.proxmox.com/?p=proxmox-backup.git;a=summary
|
||||||
Homepage: https://www.proxmox.com
|
Homepage: https://www.proxmox.com
|
||||||
|
Rules-Requires-Root: binary-targets
|
||||||
|
|
||||||
Package: proxmox-backup-server
|
Package: proxmox-backup-server
|
||||||
Architecture: any
|
Architecture: any
|
||||||
Depends: fonts-font-awesome,
|
Depends: fonts-font-awesome,
|
||||||
libjs-extjs (>= 6.0.1),
|
libjs-extjs (>= 7~),
|
||||||
libjs-qrcodejs (>= 1.20201119),
|
libjs-qrcodejs (>= 1.20201119),
|
||||||
|
libproxmox-acme-plugins,
|
||||||
libsgutils2-2,
|
libsgutils2-2,
|
||||||
libzstd1 (>= 1.3.8),
|
libzstd1 (>= 1.3.8),
|
||||||
lvm2,
|
lvm2,
|
||||||
@ -119,7 +150,7 @@ Depends: fonts-font-awesome,
|
|||||||
postfix | mail-transport-agent,
|
postfix | mail-transport-agent,
|
||||||
proxmox-backup-docs,
|
proxmox-backup-docs,
|
||||||
proxmox-mini-journalreader,
|
proxmox-mini-journalreader,
|
||||||
proxmox-widget-toolkit (>= 2.5-1),
|
proxmox-widget-toolkit (>= 3.4-3),
|
||||||
pve-xtermjs (>= 4.7.0-1),
|
pve-xtermjs (>= 4.7.0-1),
|
||||||
sg3-utils,
|
sg3-utils,
|
||||||
smartmontools,
|
smartmontools,
|
||||||
@ -143,7 +174,8 @@ Description: Proxmox Backup Client tools
|
|||||||
Package: proxmox-backup-docs
|
Package: proxmox-backup-docs
|
||||||
Build-Profiles: <!nodoc>
|
Build-Profiles: <!nodoc>
|
||||||
Section: doc
|
Section: doc
|
||||||
Depends: libjs-extjs,
|
Depends: fonts-font-awesome,
|
||||||
|
libjs-extjs,
|
||||||
libjs-mathjax,
|
libjs-mathjax,
|
||||||
${misc:Depends},
|
${misc:Depends},
|
||||||
Architecture: all
|
Architecture: all
|
||||||
@ -156,6 +188,7 @@ Depends: ${misc:Depends},
|
|||||||
${shlibs:Depends},
|
${shlibs:Depends},
|
||||||
Recommends: pve-qemu-kvm (>= 5.0.0-9),
|
Recommends: pve-qemu-kvm (>= 5.0.0-9),
|
||||||
proxmox-backup-restore-image,
|
proxmox-backup-restore-image,
|
||||||
|
Breaks: proxmox-backup-restore-image (<< 0.3.1)
|
||||||
Description: Proxmox Backup single file restore tools for pxar and block device backups
|
Description: Proxmox Backup single file restore tools for pxar and block device backups
|
||||||
This package contains the Proxmox Backup single file restore client for
|
This package contains the Proxmox Backup single file restore client for
|
||||||
restoring individual files and folders from both host/container and VM/block
|
restoring individual files and folders from both host/container and VM/block
|
||||||
|
54
debian/control.in
vendored
54
debian/control.in
vendored
@ -1,54 +0,0 @@
|
|||||||
Package: proxmox-backup-server
|
|
||||||
Architecture: any
|
|
||||||
Depends: fonts-font-awesome,
|
|
||||||
libjs-extjs (>= 6.0.1),
|
|
||||||
libjs-qrcodejs (>= 1.20201119),
|
|
||||||
libsgutils2-2,
|
|
||||||
libzstd1 (>= 1.3.8),
|
|
||||||
lvm2,
|
|
||||||
openssh-server,
|
|
||||||
pbs-i18n,
|
|
||||||
postfix | mail-transport-agent,
|
|
||||||
proxmox-backup-docs,
|
|
||||||
proxmox-mini-journalreader,
|
|
||||||
proxmox-widget-toolkit (>= 2.5-1),
|
|
||||||
pve-xtermjs (>= 4.7.0-1),
|
|
||||||
sg3-utils,
|
|
||||||
smartmontools,
|
|
||||||
${misc:Depends},
|
|
||||||
${shlibs:Depends},
|
|
||||||
Recommends: zfsutils-linux,
|
|
||||||
ifupdown2,
|
|
||||||
Description: Proxmox Backup Server daemon with tools and GUI
|
|
||||||
This package contains the Proxmox Backup Server daemons and related
|
|
||||||
tools. This includes a web-based graphical user interface.
|
|
||||||
|
|
||||||
Package: proxmox-backup-client
|
|
||||||
Architecture: any
|
|
||||||
Depends: qrencode,
|
|
||||||
${misc:Depends},
|
|
||||||
${shlibs:Depends},
|
|
||||||
Description: Proxmox Backup Client tools
|
|
||||||
This package contains the Proxmox Backup client, which provides a
|
|
||||||
simple command line tool to create and restore backups.
|
|
||||||
|
|
||||||
Package: proxmox-backup-docs
|
|
||||||
Build-Profiles: <!nodoc>
|
|
||||||
Section: doc
|
|
||||||
Depends: libjs-extjs,
|
|
||||||
libjs-mathjax,
|
|
||||||
${misc:Depends},
|
|
||||||
Architecture: all
|
|
||||||
Description: Proxmox Backup Documentation
|
|
||||||
This package contains the Proxmox Backup Documentation files.
|
|
||||||
|
|
||||||
Package: proxmox-backup-file-restore
|
|
||||||
Architecture: any
|
|
||||||
Depends: ${misc:Depends},
|
|
||||||
${shlibs:Depends},
|
|
||||||
Recommends: pve-qemu-kvm (>= 5.0.0-9),
|
|
||||||
proxmox-backup-restore-image,
|
|
||||||
Description: Proxmox Backup single file restore tools for pxar and block device backups
|
|
||||||
This package contains the Proxmox Backup single file restore client for
|
|
||||||
restoring individual files and folders from both host/container and VM/block
|
|
||||||
device backups. It includes a block device restore driver using QEMU.
|
|
42
debian/debcargo.toml
vendored
42
debian/debcargo.toml
vendored
@ -1,42 +0,0 @@
|
|||||||
overlay = "."
|
|
||||||
crate_src_path = ".."
|
|
||||||
whitelist = ["tests/*.c"]
|
|
||||||
|
|
||||||
maintainer = "Proxmox Support Team <support@proxmox.com>"
|
|
||||||
|
|
||||||
[source]
|
|
||||||
vcs_git = "git://git.proxmox.com/git/proxmox-backup.git"
|
|
||||||
vcs_browser = "https://git.proxmox.com/?p=proxmox-backup.git;a=summary"
|
|
||||||
section = "admin"
|
|
||||||
build_depends = [
|
|
||||||
"bash-completion",
|
|
||||||
"debhelper (>= 12~)",
|
|
||||||
"fonts-dejavu-core <!nodoc>",
|
|
||||||
"fonts-lato <!nodoc>",
|
|
||||||
"fonts-open-sans <!nodoc>",
|
|
||||||
"graphviz <!nodoc>",
|
|
||||||
"latexmk <!nodoc>",
|
|
||||||
"patchelf",
|
|
||||||
"pve-eslint (>= 7.18.0-1)",
|
|
||||||
"python3-docutils",
|
|
||||||
"python3-pygments",
|
|
||||||
"python3-sphinx <!nodoc>",
|
|
||||||
"rsync",
|
|
||||||
"texlive-fonts-extra <!nodoc>",
|
|
||||||
"texlive-fonts-recommended <!nodoc>",
|
|
||||||
"texlive-xetex <!nodoc>",
|
|
||||||
"xindy <!nodoc>",
|
|
||||||
]
|
|
||||||
|
|
||||||
build_depends_excludes = [
|
|
||||||
"debhelper (>=11)",
|
|
||||||
]
|
|
||||||
|
|
||||||
[packages.lib]
|
|
||||||
depends = [
|
|
||||||
"libacl1-dev",
|
|
||||||
"libfuse3-dev",
|
|
||||||
"libsystemd-dev",
|
|
||||||
"uuid-dev",
|
|
||||||
"libsgutils2-dev",
|
|
||||||
]
|
|
67
debian/postinst
vendored
67
debian/postinst
vendored
@ -4,6 +4,14 @@ set -e
|
|||||||
|
|
||||||
#DEBHELPER#
|
#DEBHELPER#
|
||||||
|
|
||||||
|
update_sync_job() {
|
||||||
|
job="$1"
|
||||||
|
|
||||||
|
echo "Updating sync job '$job' to make old 'remove-vanished' default explicit.."
|
||||||
|
proxmox-backup-manager sync-job update "$job" --remove-vanished true \
|
||||||
|
|| echo "Failed, please check sync.cfg manually!"
|
||||||
|
}
|
||||||
|
|
||||||
case "$1" in
|
case "$1" in
|
||||||
configure)
|
configure)
|
||||||
# need to have user backup in the tape group
|
# need to have user backup in the tape group
|
||||||
@ -26,48 +34,35 @@ case "$1" in
|
|||||||
fi
|
fi
|
||||||
deb-systemd-invoke $_dh_action proxmox-backup.service proxmox-backup-proxy.service >/dev/null || true
|
deb-systemd-invoke $_dh_action proxmox-backup.service proxmox-backup-proxy.service >/dev/null || true
|
||||||
|
|
||||||
# FIXME: Remove with 1.1
|
|
||||||
if test -n "$2"; then
|
if test -n "$2"; then
|
||||||
if dpkg --compare-versions "$2" 'lt' '0.9.4-1'; then
|
|
||||||
if grep -s -q -P -e '^\s+verify-schedule ' /etc/proxmox-backup/datastore.cfg; then
|
|
||||||
echo "NOTE: drop all verify schedules from datastore config."
|
|
||||||
echo "You can now add more flexible verify jobs"
|
|
||||||
flock -w 30 /etc/proxmox-backup/.datastore.lck \
|
|
||||||
sed -i '/^\s\+verify-schedule /d' /etc/proxmox-backup/datastore.cfg || true
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
if dpkg --compare-versions "$2" 'le' '0.9.5-1'; then
|
|
||||||
chown --quiet backup:backup /var/log/proxmox-backup/api/auth.log || true
|
|
||||||
fi
|
|
||||||
if dpkg --compare-versions "$2" 'le' '0.9.7-1'; then
|
|
||||||
if [ -e /etc/proxmox-backup/remote.cfg ]; then
|
|
||||||
echo "NOTE: Switching over remote.cfg to new field names.."
|
|
||||||
flock -w 30 /etc/proxmox-backup/.remote.lck \
|
|
||||||
sed -i \
|
|
||||||
-e 's/^\s\+userid /\tauth-id /g' \
|
|
||||||
/etc/proxmox-backup/remote.cfg || true
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
if dpkg --compare-versions "$2" 'le' '1.0.14-1'; then
|
|
||||||
# FIXME: Remove with 2.0
|
|
||||||
if grep -s -q -P -e '^linux:' /etc/proxmox-backup/tape.cfg; then
|
|
||||||
echo "========="
|
|
||||||
echo "= NOTE: You have now unsupported 'linux' tape drives configured."
|
|
||||||
echo "= * Execute 'udevadm control --reload-rules && udevadm trigger' to update /dev"
|
|
||||||
echo "= * Edit '/etc/proxmox-backup/tape.cfg', remove 'linux' entries and re-add over CLI/GUI"
|
|
||||||
echo "========="
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
# FIXME: remove with 2.0
|
|
||||||
if [ -d "/var/lib/proxmox-backup/tape" ] &&
|
|
||||||
[ "$(stat --printf '%a' '/var/lib/proxmox-backup/tape')" != "750" ]; then
|
|
||||||
chmod 0750 /var/lib/proxmox-backup/tape || true
|
|
||||||
fi
|
|
||||||
# FIXME: Remove in future version once we're sure no broken entries remain in anyone's files
|
# FIXME: Remove in future version once we're sure no broken entries remain in anyone's files
|
||||||
if grep -q -e ':termproxy::[^@]\+: ' /var/log/proxmox-backup/tasks/active; then
|
if grep -q -e ':termproxy::[^@]\+: ' /var/log/proxmox-backup/tasks/active; then
|
||||||
echo "Fixing up termproxy user id in task log..."
|
echo "Fixing up termproxy user id in task log..."
|
||||||
flock -w 30 /var/log/proxmox-backup/tasks/active.lock sed -i 's/:termproxy::\([^@]\+\): /:termproxy::\1@pam: /' /var/log/proxmox-backup/tasks/active || true
|
flock -w 30 /var/log/proxmox-backup/tasks/active.lock sed -i 's/:termproxy::\([^@]\+\): /:termproxy::\1@pam: /' /var/log/proxmox-backup/tasks/active || true
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
if dpkg --compare-versions "$2" 'lt' '7.1-1' && test -e /etc/proxmox-backup/sync.cfg; then
|
||||||
|
prev_job=""
|
||||||
|
|
||||||
|
# read from HERE doc because POSIX sh limitations
|
||||||
|
while read -r key value; do
|
||||||
|
if test "$key" = "sync:"; then
|
||||||
|
if test -n "$prev_job"; then
|
||||||
|
# previous job doesn't have an explicit value
|
||||||
|
update_sync_job "$prev_job"
|
||||||
|
fi
|
||||||
|
prev_job=$value
|
||||||
|
else
|
||||||
|
prev_job=""
|
||||||
|
fi
|
||||||
|
done <<EOF
|
||||||
|
$(grep -e '^sync:' -e 'remove-vanished' /etc/proxmox-backup/sync.cfg)
|
||||||
|
EOF
|
||||||
|
if test -n "$prev_job"; then
|
||||||
|
# last job doesn't have an explicit value
|
||||||
|
update_sync_job "$prev_job"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
|
|
||||||
|
8
debian/proxmox-backup-debug.bc
vendored
Normal file
8
debian/proxmox-backup-debug.bc
vendored
Normal file
@ -0,0 +1,8 @@
|
|||||||
|
# proxmox-backup-debug bash completion
|
||||||
|
|
||||||
|
# see http://tiswww.case.edu/php/chet/bash/FAQ
|
||||||
|
# and __ltrim_colon_completions() in /usr/share/bash-completion/bash_completion
|
||||||
|
# this modifies global var, but I found no better way
|
||||||
|
COMP_WORDBREAKS=${COMP_WORDBREAKS//:}
|
||||||
|
|
||||||
|
complete -C 'proxmox-backup-debug bashcomplete' proxmox-backup-debug
|
1
debian/proxmox-backup-docs.links
vendored
1
debian/proxmox-backup-docs.links
vendored
@ -1,5 +1,6 @@
|
|||||||
/usr/share/doc/proxmox-backup/proxmox-backup.pdf /usr/share/doc/proxmox-backup/html/proxmox-backup.pdf
|
/usr/share/doc/proxmox-backup/proxmox-backup.pdf /usr/share/doc/proxmox-backup/html/proxmox-backup.pdf
|
||||||
/usr/share/javascript/extjs /usr/share/doc/proxmox-backup/html/prune-simulator/extjs
|
/usr/share/javascript/extjs /usr/share/doc/proxmox-backup/html/prune-simulator/extjs
|
||||||
/usr/share/javascript/extjs /usr/share/doc/proxmox-backup/html/lto-barcode/extjs
|
/usr/share/javascript/extjs /usr/share/doc/proxmox-backup/html/lto-barcode/extjs
|
||||||
|
/usr/share/fonts-font-awesome/ /usr/share/doc/proxmox-backup/html/lto-barcode/font-awesome
|
||||||
/usr/share/javascript/extjs /usr/share/doc/proxmox-backup/html/api-viewer/extjs
|
/usr/share/javascript/extjs /usr/share/doc/proxmox-backup/html/api-viewer/extjs
|
||||||
/usr/share/javascript/mathjax /usr/share/doc/proxmox-backup/html/_static/mathjax
|
/usr/share/javascript/mathjax /usr/share/doc/proxmox-backup/html/_static/mathjax
|
||||||
|
12
debian/proxmox-backup-file-restore.postinst
vendored
12
debian/proxmox-backup-file-restore.postinst
vendored
@ -6,6 +6,7 @@ update_initramfs() {
|
|||||||
# regenerate initramfs for single file restore VM
|
# regenerate initramfs for single file restore VM
|
||||||
INST_PATH="/usr/lib/x86_64-linux-gnu/proxmox-backup/file-restore"
|
INST_PATH="/usr/lib/x86_64-linux-gnu/proxmox-backup/file-restore"
|
||||||
CACHE_PATH="/var/cache/proxmox-backup/file-restore-initramfs.img"
|
CACHE_PATH="/var/cache/proxmox-backup/file-restore-initramfs.img"
|
||||||
|
CACHE_PATH_DBG="/var/cache/proxmox-backup/file-restore-initramfs-debug.img"
|
||||||
|
|
||||||
# cleanup first, in case proxmox-file-restore was uninstalled since we do
|
# cleanup first, in case proxmox-file-restore was uninstalled since we do
|
||||||
# not want an unuseable image lying around
|
# not want an unuseable image lying around
|
||||||
@ -20,7 +21,7 @@ update_initramfs() {
|
|||||||
|
|
||||||
# avoid leftover temp file
|
# avoid leftover temp file
|
||||||
cleanup() {
|
cleanup() {
|
||||||
rm -f "$CACHE_PATH.tmp"
|
rm -f "$CACHE_PATH.tmp" "$CACHE_PATH_DBG.tmp"
|
||||||
}
|
}
|
||||||
trap cleanup EXIT
|
trap cleanup EXIT
|
||||||
|
|
||||||
@ -34,6 +35,15 @@ update_initramfs() {
|
|||||||
| cpio -o --format=newc -A -F "$CACHE_PATH.tmp" )
|
| cpio -o --format=newc -A -F "$CACHE_PATH.tmp" )
|
||||||
mv -f "$CACHE_PATH.tmp" "$CACHE_PATH"
|
mv -f "$CACHE_PATH.tmp" "$CACHE_PATH"
|
||||||
|
|
||||||
|
if [ -f "$INST_PATH/initramfs-debug.img" ]; then
|
||||||
|
echo "Updating file-restore debug initramfs..."
|
||||||
|
cp "$INST_PATH/initramfs-debug.img" "$CACHE_PATH_DBG.tmp"
|
||||||
|
( cd "$INST_PATH"; \
|
||||||
|
printf "./proxmox-restore-daemon" \
|
||||||
|
| cpio -o --format=newc -A -F "$CACHE_PATH_DBG.tmp" )
|
||||||
|
mv -f "$CACHE_PATH_DBG.tmp" "$CACHE_PATH_DBG"
|
||||||
|
fi
|
||||||
|
|
||||||
trap - EXIT
|
trap - EXIT
|
||||||
}
|
}
|
||||||
|
|
||||||
|
1
debian/proxmox-backup-server.bash-completion
vendored
1
debian/proxmox-backup-server.bash-completion
vendored
@ -1,4 +1,5 @@
|
|||||||
debian/proxmox-backup-manager.bc proxmox-backup-manager
|
debian/proxmox-backup-manager.bc proxmox-backup-manager
|
||||||
|
debian/proxmox-backup-debug.bc proxmox-backup-debug
|
||||||
debian/proxmox-tape.bc proxmox-tape
|
debian/proxmox-tape.bc proxmox-tape
|
||||||
debian/pmtx.bc pmtx
|
debian/pmtx.bc pmtx
|
||||||
debian/pmt.bc pmt
|
debian/pmt.bc pmt
|
||||||
|
3
debian/proxmox-backup-server.install
vendored
3
debian/proxmox-backup-server.install
vendored
@ -9,6 +9,7 @@ usr/lib/x86_64-linux-gnu/proxmox-backup/proxmox-backup-proxy
|
|||||||
usr/lib/x86_64-linux-gnu/proxmox-backup/proxmox-backup-banner
|
usr/lib/x86_64-linux-gnu/proxmox-backup/proxmox-backup-banner
|
||||||
usr/lib/x86_64-linux-gnu/proxmox-backup/proxmox-daily-update
|
usr/lib/x86_64-linux-gnu/proxmox-backup/proxmox-daily-update
|
||||||
usr/lib/x86_64-linux-gnu/proxmox-backup/sg-tape-cmd
|
usr/lib/x86_64-linux-gnu/proxmox-backup/sg-tape-cmd
|
||||||
|
usr/sbin/proxmox-backup-debug
|
||||||
usr/sbin/proxmox-backup-manager
|
usr/sbin/proxmox-backup-manager
|
||||||
usr/bin/pmtx
|
usr/bin/pmtx
|
||||||
usr/bin/pmt
|
usr/bin/pmt
|
||||||
@ -17,6 +18,7 @@ usr/share/javascript/proxmox-backup/index.hbs
|
|||||||
usr/share/javascript/proxmox-backup/css/ext6-pbs.css
|
usr/share/javascript/proxmox-backup/css/ext6-pbs.css
|
||||||
usr/share/javascript/proxmox-backup/images
|
usr/share/javascript/proxmox-backup/images
|
||||||
usr/share/javascript/proxmox-backup/js/proxmox-backup-gui.js
|
usr/share/javascript/proxmox-backup/js/proxmox-backup-gui.js
|
||||||
|
usr/share/man/man1/proxmox-backup-debug.1
|
||||||
usr/share/man/man1/proxmox-backup-manager.1
|
usr/share/man/man1/proxmox-backup-manager.1
|
||||||
usr/share/man/man1/proxmox-backup-proxy.1
|
usr/share/man/man1/proxmox-backup-proxy.1
|
||||||
usr/share/man/man1/proxmox-tape.1
|
usr/share/man/man1/proxmox-tape.1
|
||||||
@ -31,6 +33,7 @@ usr/share/man/man5/verification.cfg.5
|
|||||||
usr/share/man/man5/media-pool.cfg.5
|
usr/share/man/man5/media-pool.cfg.5
|
||||||
usr/share/man/man5/tape.cfg.5
|
usr/share/man/man5/tape.cfg.5
|
||||||
usr/share/man/man5/tape-job.cfg.5
|
usr/share/man/man5/tape-job.cfg.5
|
||||||
|
usr/share/zsh/vendor-completions/_proxmox-backup-debug
|
||||||
usr/share/zsh/vendor-completions/_proxmox-backup-manager
|
usr/share/zsh/vendor-completions/_proxmox-backup-manager
|
||||||
usr/share/zsh/vendor-completions/_proxmox-tape
|
usr/share/zsh/vendor-completions/_proxmox-tape
|
||||||
usr/share/zsh/vendor-completions/_pmtx
|
usr/share/zsh/vendor-completions/_pmtx
|
||||||
|
8
debian/rules
vendored
8
debian/rules
vendored
@ -32,6 +32,9 @@ override_dh_auto_build:
|
|||||||
override_dh_missing:
|
override_dh_missing:
|
||||||
dh_missing --fail-missing
|
dh_missing --fail-missing
|
||||||
|
|
||||||
|
override_dh_auto_test:
|
||||||
|
# ignore here to avoid rebuilding the binaries with the wrong target
|
||||||
|
|
||||||
override_dh_auto_install:
|
override_dh_auto_install:
|
||||||
dh_auto_install -- \
|
dh_auto_install -- \
|
||||||
PROXY_USER=backup \
|
PROXY_USER=backup \
|
||||||
@ -45,11 +48,6 @@ override_dh_installsystemd:
|
|||||||
override_dh_fixperms:
|
override_dh_fixperms:
|
||||||
dh_fixperms --exclude sg-tape-cmd
|
dh_fixperms --exclude sg-tape-cmd
|
||||||
|
|
||||||
# workaround https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=933541
|
|
||||||
# TODO: remove once available (Debian 11 ?)
|
|
||||||
override_dh_dwz:
|
|
||||||
dh_dwz --no-dwz-multifile
|
|
||||||
|
|
||||||
override_dh_strip:
|
override_dh_strip:
|
||||||
dh_strip
|
dh_strip
|
||||||
for exe in $$(find \
|
for exe in $$(find \
|
||||||
|
@ -5,6 +5,7 @@ GENERATED_SYNOPSIS := \
|
|||||||
proxmox-backup-client/synopsis.rst \
|
proxmox-backup-client/synopsis.rst \
|
||||||
proxmox-backup-client/catalog-shell-synopsis.rst \
|
proxmox-backup-client/catalog-shell-synopsis.rst \
|
||||||
proxmox-backup-manager/synopsis.rst \
|
proxmox-backup-manager/synopsis.rst \
|
||||||
|
proxmox-backup-debug/synopsis.rst \
|
||||||
proxmox-file-restore/synopsis.rst \
|
proxmox-file-restore/synopsis.rst \
|
||||||
pxar/synopsis.rst \
|
pxar/synopsis.rst \
|
||||||
pmtx/synopsis.rst \
|
pmtx/synopsis.rst \
|
||||||
@ -27,7 +28,8 @@ MAN1_PAGES := \
|
|||||||
proxmox-backup-proxy.1 \
|
proxmox-backup-proxy.1 \
|
||||||
proxmox-backup-client.1 \
|
proxmox-backup-client.1 \
|
||||||
proxmox-backup-manager.1 \
|
proxmox-backup-manager.1 \
|
||||||
proxmox-file-restore.1
|
proxmox-file-restore.1 \
|
||||||
|
proxmox-backup-debug.1
|
||||||
|
|
||||||
MAN5_PAGES := \
|
MAN5_PAGES := \
|
||||||
media-pool.cfg.5 \
|
media-pool.cfg.5 \
|
||||||
@ -46,23 +48,35 @@ PRUNE_SIMULATOR_FILES := \
|
|||||||
prune-simulator/clear-trigger.png \
|
prune-simulator/clear-trigger.png \
|
||||||
prune-simulator/prune-simulator.js
|
prune-simulator/prune-simulator.js
|
||||||
|
|
||||||
|
PRUNE_SIMULATOR_JS_SOURCE := \
|
||||||
|
/usr/share/javascript/proxmox-widget-toolkit-dev/Toolkit.js \
|
||||||
|
prune-simulator/prune-simulator_source.js
|
||||||
|
|
||||||
|
LTO_BARCODE_JS_SOURCE := \
|
||||||
|
/usr/share/javascript/proxmox-widget-toolkit-dev/Toolkit.js \
|
||||||
|
lto-barcode/code39.js \
|
||||||
|
lto-barcode/prefix-field.js \
|
||||||
|
lto-barcode/label-style.js \
|
||||||
|
lto-barcode/tape-type.js \
|
||||||
|
lto-barcode/paper-size.js \
|
||||||
|
lto-barcode/page-layout.js \
|
||||||
|
lto-barcode/page-calibration.js \
|
||||||
|
lto-barcode/label-list.js \
|
||||||
|
lto-barcode/label-setup.js \
|
||||||
|
lto-barcode/lto-barcode.js
|
||||||
|
|
||||||
LTO_BARCODE_FILES := \
|
LTO_BARCODE_FILES := \
|
||||||
lto-barcode/index.html \
|
lto-barcode/index.html \
|
||||||
lto-barcode/code39.js \
|
lto-barcode/lto-barcode-generator.js
|
||||||
lto-barcode/prefix-field.js \
|
|
||||||
lto-barcode/label-style.js \
|
|
||||||
lto-barcode/tape-type.js \
|
|
||||||
lto-barcode/paper-size.js \
|
|
||||||
lto-barcode/page-layout.js \
|
|
||||||
lto-barcode/page-calibration.js \
|
|
||||||
lto-barcode/label-list.js \
|
|
||||||
lto-barcode/label-setup.js \
|
|
||||||
lto-barcode/lto-barcode.js
|
|
||||||
|
|
||||||
API_VIEWER_SOURCES= \
|
API_VIEWER_SOURCES= \
|
||||||
api-viewer/index.html \
|
api-viewer/index.html \
|
||||||
api-viewer/apidoc.js
|
api-viewer/apidoc.js
|
||||||
|
|
||||||
|
API_VIEWER_FILES := \
|
||||||
|
api-viewer/apidata.js \
|
||||||
|
/usr/share/javascript/proxmox-widget-toolkit-dev/APIViewer.js \
|
||||||
|
|
||||||
# Sphinx documentation setup
|
# Sphinx documentation setup
|
||||||
SPHINXOPTS =
|
SPHINXOPTS =
|
||||||
SPHINXBUILD = sphinx-build
|
SPHINXBUILD = sphinx-build
|
||||||
@ -187,17 +201,32 @@ proxmox-file-restore/synopsis.rst: ${COMPILEDIR}/proxmox-file-restore
|
|||||||
proxmox-file-restore.1: proxmox-file-restore/man1.rst proxmox-file-restore/description.rst proxmox-file-restore/synopsis.rst
|
proxmox-file-restore.1: proxmox-file-restore/man1.rst proxmox-file-restore/description.rst proxmox-file-restore/synopsis.rst
|
||||||
rst2man $< >$@
|
rst2man $< >$@
|
||||||
|
|
||||||
|
proxmox-backup-debug/synopsis.rst: ${COMPILEDIR}/proxmox-backup-debug
|
||||||
|
${COMPILEDIR}/proxmox-backup-debug printdoc > proxmox-backup-debug/synopsis.rst
|
||||||
|
|
||||||
|
proxmox-backup-debug.1: proxmox-backup-debug/man1.rst proxmox-backup-debug/description.rst proxmox-backup-debug/synopsis.rst
|
||||||
|
rst2man $< >$@
|
||||||
|
|
||||||
.PHONY: onlinehelpinfo
|
.PHONY: onlinehelpinfo
|
||||||
onlinehelpinfo:
|
onlinehelpinfo:
|
||||||
@echo "Generating OnlineHelpInfo.js..."
|
@echo "Generating OnlineHelpInfo.js..."
|
||||||
$(SPHINXBUILD) -b proxmox-scanrefs $(ALLSPHINXOPTS) $(BUILDDIR)/scanrefs
|
$(SPHINXBUILD) -b proxmox-scanrefs -Q $(ALLSPHINXOPTS) $(BUILDDIR)/scanrefs
|
||||||
@echo "Build finished. OnlineHelpInfo.js is in $(BUILDDIR)/scanrefs."
|
@echo "Build finished. OnlineHelpInfo.js is in $(BUILDDIR)/scanrefs."
|
||||||
|
|
||||||
api-viewer/apidata.js: ${COMPILEDIR}/docgen
|
api-viewer/apidata.js: ${COMPILEDIR}/docgen
|
||||||
${COMPILEDIR}/docgen apidata.js >$@
|
${COMPILEDIR}/docgen apidata.js >$@
|
||||||
|
|
||||||
api-viewer/apidoc.js: api-viewer/apidata.js api-viewer/PBSAPI.js
|
api-viewer/apidoc.js: ${API_VIEWER_FILES}
|
||||||
cat api-viewer/apidata.js api-viewer/PBSAPI.js >$@
|
cat ${API_VIEWER_FILES} >$@.tmp
|
||||||
|
mv $@.tmp $@
|
||||||
|
|
||||||
|
prune-simulator/prune-simulator.js: ${PRUNE_SIMULATOR_JS_SOURCE}
|
||||||
|
cat ${PRUNE_SIMULATOR_JS_SOURCE} >$@.tmp
|
||||||
|
mv $@.tmp $@
|
||||||
|
|
||||||
|
lto-barcode/lto-barcode-generator.js: ${LTO_BARCODE_JS_SOURCE}
|
||||||
|
cat ${LTO_BARCODE_JS_SOURCE} >$@.tmp
|
||||||
|
mv $@.tmp $@
|
||||||
|
|
||||||
.PHONY: html
|
.PHONY: html
|
||||||
html: ${GENERATED_SYNOPSIS} images/proxmox-logo.svg custom.css conf.py ${PRUNE_SIMULATOR_FILES} ${LTO_BARCODE_FILES} ${API_VIEWER_SOURCES}
|
html: ${GENERATED_SYNOPSIS} images/proxmox-logo.svg custom.css conf.py ${PRUNE_SIMULATOR_FILES} ${LTO_BARCODE_FILES} ${API_VIEWER_SOURCES}
|
||||||
@ -228,6 +257,7 @@ epub3: ${GENERATED_SYNOPSIS}
|
|||||||
|
|
||||||
clean:
|
clean:
|
||||||
rm -r -f *~ *.1 ${BUILDDIR} ${GENERATED_SYNOPSIS} api-viewer/apidata.js
|
rm -r -f *~ *.1 ${BUILDDIR} ${GENERATED_SYNOPSIS} api-viewer/apidata.js
|
||||||
|
rm -f api-viewer/apidoc.js lto-barcode/lto-barcode-generator.js prune-simulator/prune-simulator.js
|
||||||
|
|
||||||
|
|
||||||
install_manual_pages: ${MAN1_PAGES} ${MAN5_PAGES}
|
install_manual_pages: ${MAN1_PAGES} ${MAN5_PAGES}
|
||||||
|
@ -1,511 +0,0 @@
|
|||||||
// avoid errors when running without development tools
|
|
||||||
if (!Ext.isDefined(Ext.global.console)) {
|
|
||||||
var console = {
|
|
||||||
dir: function() {},
|
|
||||||
log: function() {}
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
Ext.onReady(function() {
|
|
||||||
|
|
||||||
Ext.define('pve-param-schema', {
|
|
||||||
extend: 'Ext.data.Model',
|
|
||||||
fields: [
|
|
||||||
'name', 'type', 'typetext', 'description', 'verbose_description',
|
|
||||||
'enum', 'minimum', 'maximum', 'minLength', 'maxLength',
|
|
||||||
'pattern', 'title', 'requires', 'format', 'default',
|
|
||||||
'disallow', 'extends', 'links',
|
|
||||||
{
|
|
||||||
name: 'optional',
|
|
||||||
type: 'boolean'
|
|
||||||
}
|
|
||||||
]
|
|
||||||
});
|
|
||||||
|
|
||||||
var store = Ext.define('pve-updated-treestore', {
|
|
||||||
extend: 'Ext.data.TreeStore',
|
|
||||||
model: Ext.define('pve-api-doc', {
|
|
||||||
extend: 'Ext.data.Model',
|
|
||||||
fields: [
|
|
||||||
'path', 'info', 'text',
|
|
||||||
]
|
|
||||||
}),
|
|
||||||
proxy: {
|
|
||||||
type: 'memory',
|
|
||||||
data: pbsapi
|
|
||||||
},
|
|
||||||
sorters: [{
|
|
||||||
property: 'leaf',
|
|
||||||
direction: 'ASC'
|
|
||||||
}, {
|
|
||||||
property: 'text',
|
|
||||||
direction: 'ASC'
|
|
||||||
}],
|
|
||||||
filterer: 'bottomup',
|
|
||||||
doFilter: function(node) {
|
|
||||||
this.filterNodes(node, this.getFilters().getFilterFn(), true);
|
|
||||||
},
|
|
||||||
|
|
||||||
filterNodes: function(node, filterFn, parentVisible) {
|
|
||||||
var me = this,
|
|
||||||
bottomUpFiltering = me.filterer === 'bottomup',
|
|
||||||
match = filterFn(node) && parentVisible || (node.isRoot() && !me.getRootVisible()),
|
|
||||||
childNodes = node.childNodes,
|
|
||||||
len = childNodes && childNodes.length, i, matchingChildren;
|
|
||||||
|
|
||||||
if (len) {
|
|
||||||
for (i = 0; i < len; ++i) {
|
|
||||||
matchingChildren = me.filterNodes(childNodes[i], filterFn, match || bottomUpFiltering) || matchingChildren;
|
|
||||||
}
|
|
||||||
if (bottomUpFiltering) {
|
|
||||||
match = matchingChildren || match;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
node.set("visible", match, me._silentOptions);
|
|
||||||
return match;
|
|
||||||
},
|
|
||||||
|
|
||||||
}).create();
|
|
||||||
|
|
||||||
var render_description = function(value, metaData, record) {
|
|
||||||
var pdef = record.data;
|
|
||||||
|
|
||||||
value = pdef.verbose_description || value;
|
|
||||||
|
|
||||||
// TODO: try to render asciidoc correctly
|
|
||||||
|
|
||||||
metaData.style = 'white-space:pre-wrap;'
|
|
||||||
|
|
||||||
return Ext.htmlEncode(value);
|
|
||||||
};
|
|
||||||
|
|
||||||
var render_type = function(value, metaData, record) {
|
|
||||||
var pdef = record.data;
|
|
||||||
|
|
||||||
return pdef['enum'] ? 'enum' : (pdef.type || 'string');
|
|
||||||
};
|
|
||||||
|
|
||||||
var render_format = function(value, metaData, record) {
|
|
||||||
var pdef = record.data;
|
|
||||||
|
|
||||||
metaData.style = 'white-space:normal;'
|
|
||||||
|
|
||||||
if (pdef.typetext)
|
|
||||||
return Ext.htmlEncode(pdef.typetext);
|
|
||||||
|
|
||||||
if (pdef['enum'])
|
|
||||||
return pdef['enum'].join(' | ');
|
|
||||||
|
|
||||||
if (pdef.format)
|
|
||||||
return pdef.format;
|
|
||||||
|
|
||||||
if (pdef.pattern)
|
|
||||||
return Ext.htmlEncode(pdef.pattern);
|
|
||||||
|
|
||||||
return '';
|
|
||||||
};
|
|
||||||
|
|
||||||
var real_path = function(path) {
|
|
||||||
return path.replace(/^.*\/_upgrade_(\/)?/, "/");
|
|
||||||
};
|
|
||||||
|
|
||||||
var permission_text = function(permission) {
|
|
||||||
let permhtml = "";
|
|
||||||
|
|
||||||
if (permission.user) {
|
|
||||||
if (!permission.description) {
|
|
||||||
if (permission.user === 'world') {
|
|
||||||
permhtml += "Accessible without any authentication.";
|
|
||||||
} else if (permission.user === 'all') {
|
|
||||||
permhtml += "Accessible by all authenticated users.";
|
|
||||||
} else {
|
|
||||||
permhtml += 'Onyl accessible by user "' +
|
|
||||||
permission.user + '"';
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else if (permission.check) {
|
|
||||||
permhtml += "<pre>Check: " +
|
|
||||||
Ext.htmlEncode(Ext.JSON.encode(permission.check)) + "</pre>";
|
|
||||||
} else if (permission.userParam) {
|
|
||||||
permhtml += `<div>Check if user matches parameter '${permission.userParam}'`;
|
|
||||||
} else if (permission.or) {
|
|
||||||
permhtml += "<div>Or<div style='padding-left: 10px;'>";
|
|
||||||
Ext.Array.each(permission.or, function(sub_permission) {
|
|
||||||
permhtml += permission_text(sub_permission);
|
|
||||||
})
|
|
||||||
permhtml += "</div></div>";
|
|
||||||
} else if (permission.and) {
|
|
||||||
permhtml += "<div>And<div style='padding-left: 10px;'>";
|
|
||||||
Ext.Array.each(permission.and, function(sub_permission) {
|
|
||||||
permhtml += permission_text(sub_permission);
|
|
||||||
})
|
|
||||||
permhtml += "</div></div>";
|
|
||||||
} else {
|
|
||||||
//console.log(permission);
|
|
||||||
permhtml += "Unknown syntax!";
|
|
||||||
}
|
|
||||||
|
|
||||||
return permhtml;
|
|
||||||
};
|
|
||||||
|
|
||||||
var render_docu = function(data) {
|
|
||||||
var md = data.info;
|
|
||||||
|
|
||||||
// console.dir(data);
|
|
||||||
|
|
||||||
var items = [];
|
|
||||||
|
|
||||||
var clicmdhash = {
|
|
||||||
GET: 'get',
|
|
||||||
POST: 'create',
|
|
||||||
PUT: 'set',
|
|
||||||
DELETE: 'delete'
|
|
||||||
};
|
|
||||||
|
|
||||||
Ext.Array.each(['GET', 'POST', 'PUT', 'DELETE'], function(method) {
|
|
||||||
var info = md[method];
|
|
||||||
if (info) {
|
|
||||||
|
|
||||||
var usage = "";
|
|
||||||
|
|
||||||
usage += "<table><tr><td>HTTP: </td><td>"
|
|
||||||
+ method + " " + real_path("/api2/json" + data.path) + "</td></tr>";
|
|
||||||
|
|
||||||
var sections = [
|
|
||||||
{
|
|
||||||
title: 'Description',
|
|
||||||
html: Ext.htmlEncode(info.description),
|
|
||||||
bodyPadding: 10
|
|
||||||
},
|
|
||||||
{
|
|
||||||
title: 'Usage',
|
|
||||||
html: usage,
|
|
||||||
bodyPadding: 10
|
|
||||||
}
|
|
||||||
];
|
|
||||||
|
|
||||||
if (info.parameters && info.parameters.properties) {
|
|
||||||
|
|
||||||
var pstore = Ext.create('Ext.data.Store', {
|
|
||||||
model: 'pve-param-schema',
|
|
||||||
proxy: {
|
|
||||||
type: 'memory'
|
|
||||||
},
|
|
||||||
groupField: 'optional',
|
|
||||||
sorters: [
|
|
||||||
{
|
|
||||||
property: 'name',
|
|
||||||
direction: 'ASC'
|
|
||||||
}
|
|
||||||
]
|
|
||||||
});
|
|
||||||
|
|
||||||
Ext.Object.each(info.parameters.properties, function(name, pdef) {
|
|
||||||
pdef.name = name;
|
|
||||||
pstore.add(pdef);
|
|
||||||
});
|
|
||||||
|
|
||||||
pstore.sort();
|
|
||||||
|
|
||||||
var groupingFeature = Ext.create('Ext.grid.feature.Grouping',{
|
|
||||||
enableGroupingMenu: false,
|
|
||||||
groupHeaderTpl: '<tpl if="groupValue">Optional</tpl><tpl if="!groupValue">Required</tpl>'
|
|
||||||
});
|
|
||||||
|
|
||||||
sections.push({
|
|
||||||
xtype: 'gridpanel',
|
|
||||||
title: 'Parameters',
|
|
||||||
features: [groupingFeature],
|
|
||||||
store: pstore,
|
|
||||||
viewConfig: {
|
|
||||||
trackOver: false,
|
|
||||||
stripeRows: true
|
|
||||||
},
|
|
||||||
columns: [
|
|
||||||
{
|
|
||||||
header: 'Name',
|
|
||||||
dataIndex: 'name',
|
|
||||||
flex: 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
header: 'Type',
|
|
||||||
dataIndex: 'type',
|
|
||||||
renderer: render_type,
|
|
||||||
flex: 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
header: 'Default',
|
|
||||||
dataIndex: 'default',
|
|
||||||
flex: 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
header: 'Format',
|
|
||||||
dataIndex: 'type',
|
|
||||||
renderer: render_format,
|
|
||||||
flex: 2
|
|
||||||
},
|
|
||||||
{
|
|
||||||
header: 'Description',
|
|
||||||
dataIndex: 'description',
|
|
||||||
renderer: render_description,
|
|
||||||
flex: 6
|
|
||||||
}
|
|
||||||
]
|
|
||||||
});
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
if (info.returns) {
|
|
||||||
|
|
||||||
var retinf = info.returns;
|
|
||||||
var rtype = retinf.type;
|
|
||||||
if (!rtype && retinf.items)
|
|
||||||
rtype = 'array';
|
|
||||||
if (!rtype)
|
|
||||||
rtype = 'object';
|
|
||||||
|
|
||||||
var rpstore = Ext.create('Ext.data.Store', {
|
|
||||||
model: 'pve-param-schema',
|
|
||||||
proxy: {
|
|
||||||
type: 'memory'
|
|
||||||
},
|
|
||||||
groupField: 'optional',
|
|
||||||
sorters: [
|
|
||||||
{
|
|
||||||
property: 'name',
|
|
||||||
direction: 'ASC'
|
|
||||||
}
|
|
||||||
]
|
|
||||||
});
|
|
||||||
|
|
||||||
var properties;
|
|
||||||
if (rtype === 'array' && retinf.items.properties) {
|
|
||||||
properties = retinf.items.properties;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (rtype === 'object' && retinf.properties) {
|
|
||||||
properties = retinf.properties;
|
|
||||||
}
|
|
||||||
|
|
||||||
Ext.Object.each(properties, function(name, pdef) {
|
|
||||||
pdef.name = name;
|
|
||||||
rpstore.add(pdef);
|
|
||||||
});
|
|
||||||
|
|
||||||
rpstore.sort();
|
|
||||||
|
|
||||||
var groupingFeature = Ext.create('Ext.grid.feature.Grouping',{
|
|
||||||
enableGroupingMenu: false,
|
|
||||||
groupHeaderTpl: '<tpl if="groupValue">Optional</tpl><tpl if="!groupValue">Obligatory</tpl>'
|
|
||||||
});
|
|
||||||
var returnhtml;
|
|
||||||
if (retinf.items) {
|
|
||||||
returnhtml = '<pre>items: ' + Ext.htmlEncode(JSON.stringify(retinf.items, null, 4)) + '</pre>';
|
|
||||||
}
|
|
||||||
|
|
||||||
if (retinf.properties) {
|
|
||||||
returnhtml = returnhtml || '';
|
|
||||||
returnhtml += '<pre>properties:' + Ext.htmlEncode(JSON.stringify(retinf.properties, null, 4)) + '</pre>';
|
|
||||||
}
|
|
||||||
|
|
||||||
var rawSection = Ext.create('Ext.panel.Panel', {
|
|
||||||
bodyPadding: '0px 10px 10px 10px',
|
|
||||||
html: returnhtml,
|
|
||||||
hidden: true
|
|
||||||
});
|
|
||||||
|
|
||||||
sections.push({
|
|
||||||
xtype: 'gridpanel',
|
|
||||||
title: 'Returns: ' + rtype,
|
|
||||||
features: [groupingFeature],
|
|
||||||
store: rpstore,
|
|
||||||
viewConfig: {
|
|
||||||
trackOver: false,
|
|
||||||
stripeRows: true
|
|
||||||
},
|
|
||||||
columns: [
|
|
||||||
{
|
|
||||||
header: 'Name',
|
|
||||||
dataIndex: 'name',
|
|
||||||
flex: 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
header: 'Type',
|
|
||||||
dataIndex: 'type',
|
|
||||||
renderer: render_type,
|
|
||||||
flex: 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
header: 'Default',
|
|
||||||
dataIndex: 'default',
|
|
||||||
flex: 1
|
|
||||||
},
|
|
||||||
{
|
|
||||||
header: 'Format',
|
|
||||||
dataIndex: 'type',
|
|
||||||
renderer: render_format,
|
|
||||||
flex: 2
|
|
||||||
},
|
|
||||||
{
|
|
||||||
header: 'Description',
|
|
||||||
dataIndex: 'description',
|
|
||||||
renderer: render_description,
|
|
||||||
flex: 6
|
|
||||||
}
|
|
||||||
],
|
|
||||||
bbar: [
|
|
||||||
{
|
|
||||||
xtype: 'button',
|
|
||||||
text: 'Show RAW',
|
|
||||||
handler: function(btn) {
|
|
||||||
rawSection.setVisible(!rawSection.isVisible());
|
|
||||||
btn.setText(rawSection.isVisible() ? 'Hide RAW' : 'Show RAW');
|
|
||||||
}}
|
|
||||||
]
|
|
||||||
});
|
|
||||||
|
|
||||||
sections.push(rawSection);
|
|
||||||
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!data.path.match(/\/_upgrade_/)) {
|
|
||||||
var permhtml = '';
|
|
||||||
|
|
||||||
if (!info.permissions) {
|
|
||||||
permhtml = "Root only.";
|
|
||||||
} else {
|
|
||||||
if (info.permissions.description) {
|
|
||||||
permhtml += "<div style='white-space:pre-wrap;padding-bottom:10px;'>" +
|
|
||||||
Ext.htmlEncode(info.permissions.description) + "</div>";
|
|
||||||
}
|
|
||||||
permhtml += permission_text(info.permissions);
|
|
||||||
}
|
|
||||||
|
|
||||||
// we do not have this information for PBS api
|
|
||||||
//if (!info.allowtoken) {
|
|
||||||
// permhtml += "<br />This API endpoint is not available for API tokens."
|
|
||||||
//}
|
|
||||||
|
|
||||||
sections.push({
|
|
||||||
title: 'Required permissions',
|
|
||||||
bodyPadding: 10,
|
|
||||||
html: permhtml
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
items.push({
|
|
||||||
title: method,
|
|
||||||
autoScroll: true,
|
|
||||||
defaults: {
|
|
||||||
border: false
|
|
||||||
},
|
|
||||||
items: sections
|
|
||||||
});
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
var ct = Ext.getCmp('docview');
|
|
||||||
ct.setTitle("Path: " + real_path(data.path));
|
|
||||||
ct.removeAll(true);
|
|
||||||
ct.add(items);
|
|
||||||
ct.setActiveTab(0);
|
|
||||||
};
|
|
||||||
|
|
||||||
Ext.define('Ext.form.SearchField', {
|
|
||||||
extend: 'Ext.form.field.Text',
|
|
||||||
alias: 'widget.searchfield',
|
|
||||||
|
|
||||||
emptyText: 'Search...',
|
|
||||||
|
|
||||||
flex: 1,
|
|
||||||
|
|
||||||
inputType: 'search',
|
|
||||||
listeners: {
|
|
||||||
'change': function(){
|
|
||||||
|
|
||||||
var value = this.getValue();
|
|
||||||
if (!Ext.isEmpty(value)) {
|
|
||||||
store.filter({
|
|
||||||
property: 'path',
|
|
||||||
value: value,
|
|
||||||
anyMatch: true
|
|
||||||
});
|
|
||||||
} else {
|
|
||||||
store.clearFilter();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
var tree = Ext.create('Ext.tree.Panel', {
|
|
||||||
title: 'Resource Tree',
|
|
||||||
tbar: [
|
|
||||||
{
|
|
||||||
xtype: 'searchfield',
|
|
||||||
}
|
|
||||||
],
|
|
||||||
tools: [
|
|
||||||
{
|
|
||||||
type: 'expand',
|
|
||||||
tooltip: 'Expand all',
|
|
||||||
tooltipType: 'title',
|
|
||||||
callback: (tree) => tree.expandAll(),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
type: 'collapse',
|
|
||||||
tooltip: 'Collapse all',
|
|
||||||
tooltipType: 'title',
|
|
||||||
callback: (tree) => tree.collapseAll(),
|
|
||||||
},
|
|
||||||
],
|
|
||||||
store: store,
|
|
||||||
width: 200,
|
|
||||||
region: 'west',
|
|
||||||
split: true,
|
|
||||||
margins: '5 0 5 5',
|
|
||||||
rootVisible: false,
|
|
||||||
listeners: {
|
|
||||||
selectionchange: function(v, selections) {
|
|
||||||
if (!selections[0])
|
|
||||||
return;
|
|
||||||
var rec = selections[0];
|
|
||||||
render_docu(rec.data);
|
|
||||||
location.hash = '#' + rec.data.path;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
Ext.create('Ext.container.Viewport', {
|
|
||||||
layout: 'border',
|
|
||||||
renderTo: Ext.getBody(),
|
|
||||||
items: [
|
|
||||||
tree,
|
|
||||||
{
|
|
||||||
xtype: 'tabpanel',
|
|
||||||
title: 'Documentation',
|
|
||||||
id: 'docview',
|
|
||||||
region: 'center',
|
|
||||||
margins: '5 5 5 0',
|
|
||||||
layout: 'fit',
|
|
||||||
items: []
|
|
||||||
}
|
|
||||||
]
|
|
||||||
});
|
|
||||||
|
|
||||||
var deepLink = function() {
|
|
||||||
var path = window.location.hash.substring(1).replace(/\/\s*$/, '')
|
|
||||||
var endpoint = store.findNode('path', path);
|
|
||||||
|
|
||||||
if (endpoint) {
|
|
||||||
tree.getSelectionModel().select(endpoint);
|
|
||||||
tree.expandPath(endpoint.getPath());
|
|
||||||
render_docu(endpoint.data);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
window.onhashchange = deepLink;
|
|
||||||
|
|
||||||
deepLink();
|
|
||||||
|
|
||||||
});
|
|
@ -1,31 +1,33 @@
|
|||||||
Backup Client Usage
|
Backup Client Usage
|
||||||
===================
|
===================
|
||||||
|
|
||||||
The command line client is called :command:`proxmox-backup-client`.
|
The command line client for Proxmox Backup Server is called
|
||||||
|
:command:`proxmox-backup-client`.
|
||||||
|
|
||||||
.. _client_repository:
|
.. _client_repository:
|
||||||
|
|
||||||
Backup Repository Locations
|
Backup Repository Locations
|
||||||
---------------------------
|
---------------------------
|
||||||
|
|
||||||
The client uses the following notation to specify a datastore repository
|
The client uses the following format to specify a datastore repository
|
||||||
on the backup server.
|
on the backup server (where username is specified in the form of user@realm):
|
||||||
|
|
||||||
[[username@]server[:port]:]datastore
|
[[username@]server[:port]:]datastore
|
||||||
|
|
||||||
The default value for ``username`` is ``root@pam``. If no server is specified,
|
The default value for ``username`` is ``root@pam``. If no server is specified,
|
||||||
the default is the local host (``localhost``).
|
the default is the local host (``localhost``).
|
||||||
|
|
||||||
You can specify a port if your backup server is only reachable on a different
|
You can specify a port if your backup server is only reachable on a non-default
|
||||||
port (e.g. with NAT and port forwarding).
|
port (for example, with NAT and port forwarding configurations).
|
||||||
|
|
||||||
Note that if the server is an IPv6 address, you have to write it with square
|
Note that if the server uses an IPv6 address, you have to write it with square
|
||||||
brackets (for example, `[fe80::01]`).
|
brackets (for example, `[fe80::01]`).
|
||||||
|
|
||||||
You can pass the repository with the ``--repository`` command line option, or
|
You can pass the repository with the ``--repository`` command line option, or
|
||||||
by setting the ``PBS_REPOSITORY`` environment variable.
|
by setting the ``PBS_REPOSITORY`` environment variable.
|
||||||
|
|
||||||
Here some examples of valid repositories and the real values
|
Below are some examples of valid repositories and their corresponding real
|
||||||
|
values:
|
||||||
|
|
||||||
================================ ================== ================== ===========
|
================================ ================== ================== ===========
|
||||||
Example User Host:Port Datastore
|
Example User Host:Port Datastore
|
||||||
@ -46,16 +48,31 @@ Environment Variables
|
|||||||
The default backup repository.
|
The default backup repository.
|
||||||
|
|
||||||
``PBS_PASSWORD``
|
``PBS_PASSWORD``
|
||||||
When set, this value is used for the password required for the backup server.
|
When set, this value is used as the password for the backup server.
|
||||||
You can also set this to a API token secret.
|
You can also set this to an API token secret.
|
||||||
|
|
||||||
|
``PBS_PASSWORD_FD``, ``PBS_PASSWORD_FILE``, ``PBS_PASSWORD_CMD``
|
||||||
|
Like ``PBS_PASSWORD``, but read data from an open file descriptor, a file
|
||||||
|
name or from the `stdout` of a command, respectively. The first defined
|
||||||
|
environment variable from the order above is preferred.
|
||||||
|
|
||||||
``PBS_ENCRYPTION_PASSWORD``
|
``PBS_ENCRYPTION_PASSWORD``
|
||||||
When set, this value is used to access the secret encryption key (if
|
When set, this value is used to access the secret encryption key (if
|
||||||
protected by password).
|
protected by password).
|
||||||
|
|
||||||
``PBS_FINGERPRINT`` When set, this value is used to verify the server
|
``PBS_ENCRYPTION_PASSWORD_FD``, ``PBS_ENCRYPTION_PASSWORD_FILE``, ``PBS_ENCRYPTION_PASSWORD_CMD``
|
||||||
certificate (only used if the system CA certificates cannot validate the
|
Like ``PBS_ENCRYPTION_PASSWORD``, but read data from an open file descriptor,
|
||||||
certificate).
|
a file name or from the `stdout` of a command, respectively. The first
|
||||||
|
defined environment variable from the order above is preferred.
|
||||||
|
|
||||||
|
``PBS_FINGERPRINT``
|
||||||
|
When set, this value is used to verify the server certificate (only used if
|
||||||
|
the system CA certificates cannot validate the certificate).
|
||||||
|
|
||||||
|
|
||||||
|
.. Note:: Passwords must be valid UTF-8 and may not contain newlines. For your
|
||||||
|
convienience, Proxmox Backup Server only uses the first line as password, so
|
||||||
|
you can add arbitrary comments after the first newline.
|
||||||
|
|
||||||
|
|
||||||
Output Format
|
Output Format
|
||||||
@ -70,14 +87,15 @@ Creating Backups
|
|||||||
----------------
|
----------------
|
||||||
|
|
||||||
This section explains how to create a backup from within the machine. This can
|
This section explains how to create a backup from within the machine. This can
|
||||||
be a physical host, a virtual machine, or a container. Such backups may contain file
|
be a physical host, a virtual machine, or a container. Such backups may contain
|
||||||
and image archives. There are no restrictions in this case.
|
file and image archives. There are no restrictions in this case.
|
||||||
|
|
||||||
.. note:: If you want to backup virtual machines or containers on Proxmox VE, see :ref:`pve-integration`.
|
.. Note:: If you want to backup virtual machines or containers on Proxmox VE,
|
||||||
|
see :ref:`pve-integration`.
|
||||||
|
|
||||||
For the following example you need to have a backup server set up, working
|
For the following example, you need to have a backup server set up, have working
|
||||||
credentials and need to know the repository name.
|
credentials, and know the repository name.
|
||||||
In the following examples we use ``backup-server:store1``.
|
In the following examples, we use ``backup-server:store1``.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -91,12 +109,12 @@ In the following examples we use ``backup-server:store1``.
|
|||||||
Uploaded 12129 chunks in 87 seconds (564 MB/s).
|
Uploaded 12129 chunks in 87 seconds (564 MB/s).
|
||||||
End Time: 2019-12-03T10:36:29+01:00
|
End Time: 2019-12-03T10:36:29+01:00
|
||||||
|
|
||||||
This will prompt you for a password and then uploads a file archive named
|
This will prompt you for a password, then upload a file archive named
|
||||||
``root.pxar`` containing all the files in the ``/`` directory.
|
``root.pxar`` containing all the files in the ``/`` directory.
|
||||||
|
|
||||||
.. Caution:: Please note that the proxmox-backup-client does not
|
.. Caution:: Please note that proxmox-backup-client does not
|
||||||
automatically include mount points. Instead, you will see a short
|
automatically include mount points. Instead, you will see a short
|
||||||
``skip mount point`` notice for each of them. The idea is to
|
``skip mount point`` message for each of them. The idea is to
|
||||||
create a separate file archive for each mounted disk. You can
|
create a separate file archive for each mounted disk. You can
|
||||||
explicitly include them using the ``--include-dev`` option
|
explicitly include them using the ``--include-dev`` option
|
||||||
(i.e. ``--include-dev /boot/efi``). You can use this option
|
(i.e. ``--include-dev /boot/efi``). You can use this option
|
||||||
@ -104,19 +122,19 @@ This will prompt you for a password and then uploads a file archive named
|
|||||||
|
|
||||||
The ``--repository`` option can get quite long and is used by all
|
The ``--repository`` option can get quite long and is used by all
|
||||||
commands. You can avoid having to enter this value by setting the
|
commands. You can avoid having to enter this value by setting the
|
||||||
environment variable ``PBS_REPOSITORY``. Note that if you would like this to remain set
|
environment variable ``PBS_REPOSITORY``. Note that if you would like this to
|
||||||
over multiple sessions, you should instead add the below line to your
|
remain set over multiple sessions, you should instead add the below line to your
|
||||||
``.bashrc`` file.
|
``.bashrc`` file.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
# export PBS_REPOSITORY=backup-server:store1
|
# export PBS_REPOSITORY=backup-server:store1
|
||||||
|
|
||||||
After this you can execute all commands without specifying the ``--repository``
|
After this, you can execute all commands without having to specify the
|
||||||
option.
|
``--repository`` option.
|
||||||
|
|
||||||
One single backup is allowed to contain more than one archive. For example, if
|
A single backup is allowed to contain more than one archive. For example, if
|
||||||
you want to backup two disks mounted at ``/mnt/disk1`` and ``/mnt/disk2``:
|
you want to back up two disks mounted at ``/mnt/disk1`` and ``/mnt/disk2``:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -130,26 +148,26 @@ archive source at the client. The format is:
|
|||||||
|
|
||||||
<archive-name>.<type>:<source-path>
|
<archive-name>.<type>:<source-path>
|
||||||
|
|
||||||
Common types are ``.pxar`` for file archives, and ``.img`` for block
|
Common types are ``.pxar`` for file archives and ``.img`` for block
|
||||||
device images. To create a backup of a block device run the following command:
|
device images. To create a backup of a block device, run the following command:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
# proxmox-backup-client backup mydata.img:/dev/mylvm/mydata
|
# proxmox-backup-client backup mydata.img:/dev/mylvm/mydata
|
||||||
|
|
||||||
|
|
||||||
Excluding files/folders from a backup
|
Excluding Files/Directories from a Backup
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
Sometimes it is desired to exclude certain files or folders from a backup archive.
|
Sometimes it is desired to exclude certain files or directories from a backup archive.
|
||||||
To tell the Proxmox Backup client when and how to ignore files and directories,
|
To tell the Proxmox Backup client when and how to ignore files and directories,
|
||||||
place a text file called ``.pxarexclude`` in the filesystem hierarchy.
|
place a text file named ``.pxarexclude`` in the filesystem hierarchy.
|
||||||
Whenever the backup client encounters such a file in a directory, it interprets
|
Whenever the backup client encounters such a file in a directory, it interprets
|
||||||
each line as glob match patterns for files and directories that are to be excluded
|
each line as a glob match pattern for files and directories that are to be excluded
|
||||||
from the backup.
|
from the backup.
|
||||||
|
|
||||||
The file must contain a single glob pattern per line. Empty lines are ignored.
|
The file must contain a single glob pattern per line. Empty lines and lines
|
||||||
The same is true for lines starting with ``#``, which indicates a comment.
|
starting with ``#`` (indicating a comment) are ignored.
|
||||||
A ``!`` at the beginning of a line reverses the glob match pattern from an exclusion
|
A ``!`` at the beginning of a line reverses the glob match pattern from an exclusion
|
||||||
to an explicit inclusion. This makes it possible to exclude all entries in a
|
to an explicit inclusion. This makes it possible to exclude all entries in a
|
||||||
directory except for a few single files/subdirectories.
|
directory except for a few single files/subdirectories.
|
||||||
@ -160,23 +178,24 @@ the given patterns. It is only possible to match files in this directory and its
|
|||||||
``\`` is used to escape special glob characters.
|
``\`` is used to escape special glob characters.
|
||||||
``?`` matches any single character.
|
``?`` matches any single character.
|
||||||
``*`` matches any character, including an empty string.
|
``*`` matches any character, including an empty string.
|
||||||
``**`` is used to match subdirectories. It can be used to, for example, exclude
|
``**`` is used to match current directory and subdirectories. For example, with
|
||||||
all files ending in ``.tmp`` within the directory or subdirectories with the
|
the pattern ``**/*.tmp``, it would exclude all files ending in ``.tmp`` within
|
||||||
following pattern ``**/*.tmp``.
|
a directory and its subdirectories.
|
||||||
``[...]`` matches a single character from any of the provided characters within
|
``[...]`` matches a single character from any of the provided characters within
|
||||||
the brackets. ``[!...]`` does the complementary and matches any single character
|
the brackets. ``[!...]`` does the complementary and matches any single character
|
||||||
not contained within the brackets. It is also possible to specify ranges with two
|
not contained within the brackets. It is also possible to specify ranges with two
|
||||||
characters separated by ``-``. For example, ``[a-z]`` matches any lowercase
|
characters separated by ``-``. For example, ``[a-z]`` matches any lowercase
|
||||||
alphabetic character and ``[0-9]`` matches any one single digit.
|
alphabetic character, and ``[0-9]`` matches any single digit.
|
||||||
|
|
||||||
The order of the glob match patterns defines whether a file is included or
|
The order of the glob match patterns defines whether a file is included or
|
||||||
excluded, that is to say later entries override previous ones.
|
excluded, that is to say, later entries override earlier ones.
|
||||||
This is also true for match patterns encountered deeper down the directory tree,
|
This is also true for match patterns encountered deeper down the directory tree,
|
||||||
which can override a previous exclusion.
|
which can override a previous exclusion.
|
||||||
Be aware that excluded directories will **not** be read by the backup client.
|
|
||||||
Thus, a ``.pxarexclude`` file in an excluded subdirectory will have no effect.
|
.. Note:: Excluded directories will **not** be read by the backup client. Thus,
|
||||||
``.pxarexclude`` files are treated as regular files and will be included in the
|
a ``.pxarexclude`` file in an excluded subdirectory will have no effect.
|
||||||
backup archive.
|
``.pxarexclude`` files are treated as regular files and will be included in
|
||||||
|
the backup archive.
|
||||||
|
|
||||||
For example, consider the following directory structure:
|
For example, consider the following directory structure:
|
||||||
|
|
||||||
@ -264,7 +283,7 @@ You can avoid entering the passwords by setting the environment
|
|||||||
variables ``PBS_PASSWORD`` and ``PBS_ENCRYPTION_PASSWORD``.
|
variables ``PBS_PASSWORD`` and ``PBS_ENCRYPTION_PASSWORD``.
|
||||||
|
|
||||||
|
|
||||||
Using a master key to store and recover encryption keys
|
Using a Master Key to Store and Recover Encryption Keys
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
You can also use ``proxmox-backup-client key`` to create an RSA public/private
|
You can also use ``proxmox-backup-client key`` to create an RSA public/private
|
||||||
@ -344,7 +363,7 @@ To set up a master key:
|
|||||||
keep keys ordered and in a place that is separate from the contents being
|
keep keys ordered and in a place that is separate from the contents being
|
||||||
backed up. It can happen, for example, that you back up an entire system, using
|
backed up. It can happen, for example, that you back up an entire system, using
|
||||||
a key on that system. If the system then becomes inaccessible for any reason
|
a key on that system. If the system then becomes inaccessible for any reason
|
||||||
and needs to be restored, this will not be possible as the encryption key will be
|
and needs to be restored, this will not be possible, as the encryption key will be
|
||||||
lost along with the broken system.
|
lost along with the broken system.
|
||||||
|
|
||||||
It is recommended that you keep your master key safe, but easily accessible, in
|
It is recommended that you keep your master key safe, but easily accessible, in
|
||||||
@ -366,10 +385,10 @@ version of your master key. The following command sends the output of the
|
|||||||
Restoring Data
|
Restoring Data
|
||||||
--------------
|
--------------
|
||||||
|
|
||||||
The regular creation of backups is a necessary step to avoiding data
|
The regular creation of backups is a necessary step in avoiding data loss. More
|
||||||
loss. More importantly, however, is the restoration. It is good practice to perform
|
importantly, however, is the restoration. It is good practice to perform
|
||||||
periodic recovery tests to ensure that you can access the data in
|
periodic recovery tests to ensure that you can access the data in case of
|
||||||
case of problems.
|
disaster.
|
||||||
|
|
||||||
First, you need to find the snapshot which you want to restore. The snapshot
|
First, you need to find the snapshot which you want to restore. The snapshot
|
||||||
list command provides a list of all the snapshots on the server:
|
list command provides a list of all the snapshots on the server:
|
||||||
@ -428,23 +447,22 @@ to use the interactive recovery shell.
|
|||||||
|
|
||||||
The interactive recovery shell is a minimal command line interface that
|
The interactive recovery shell is a minimal command line interface that
|
||||||
utilizes the metadata stored in the catalog to quickly list, navigate and
|
utilizes the metadata stored in the catalog to quickly list, navigate and
|
||||||
search files in a file archive.
|
search for files in a file archive.
|
||||||
To restore files, you can select them individually or match them with a glob
|
To restore files, you can select them individually or match them with a glob
|
||||||
pattern.
|
pattern.
|
||||||
|
|
||||||
Using the catalog for navigation reduces the overhead considerably because only
|
Using the catalog for navigation reduces the overhead considerably because only
|
||||||
the catalog needs to be downloaded and, optionally, decrypted.
|
the catalog needs to be downloaded and, optionally, decrypted.
|
||||||
The actual chunks are only accessed if the metadata in the catalog is not enough
|
The actual chunks are only accessed if the metadata in the catalog is
|
||||||
or for the actual restore.
|
insufficient or for the actual restore.
|
||||||
|
|
||||||
Similar to common UNIX shells ``cd`` and ``ls`` are the commands used to change
|
Similar to common UNIX shells, ``cd`` and ``ls`` are the commands used to change
|
||||||
working directory and list directory contents in the archive.
|
working directory and list directory contents in the archive.
|
||||||
``pwd`` shows the full path of the current working directory with respect to the
|
``pwd`` shows the full path of the current working directory with respect to the
|
||||||
archive root.
|
archive root.
|
||||||
|
|
||||||
Being able to quickly search the contents of the archive is a commonly needed feature.
|
The ability to quickly search the contents of the archive is a commonly required
|
||||||
That's where the catalog is most valuable.
|
feature. That's where the catalog is most valuable. For example:
|
||||||
For example:
|
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -455,8 +473,8 @@ For example:
|
|||||||
pxar:/ > restore-selected /target/path
|
pxar:/ > restore-selected /target/path
|
||||||
...
|
...
|
||||||
|
|
||||||
This will find and print all files ending in ``.txt`` located in ``etc/`` or a
|
This will find and print all files ending in ``.txt`` located in ``etc/`` or its
|
||||||
subdirectory and add the corresponding pattern to the list for subsequent restores.
|
subdirectories, and add the corresponding pattern to the list for subsequent restores.
|
||||||
``list-selected`` shows these patterns and ``restore-selected`` finally restores
|
``list-selected`` shows these patterns and ``restore-selected`` finally restores
|
||||||
all files in the archive matching the patterns to ``/target/path`` on the local
|
all files in the archive matching the patterns to ``/target/path`` on the local
|
||||||
host. This will scan the whole archive.
|
host. This will scan the whole archive.
|
||||||
@ -481,7 +499,7 @@ Mounting of Archives via FUSE
|
|||||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
The :term:`FUSE` implementation for the pxar archive allows you to mount a
|
The :term:`FUSE` implementation for the pxar archive allows you to mount a
|
||||||
file archive as a read-only filesystem to a mountpoint on your host.
|
file archive as a read-only filesystem to a mount point on your host.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -497,7 +515,7 @@ This allows you to access the full contents of the archive in a seamless manner.
|
|||||||
load on your host, depending on the operations you perform on the mounted
|
load on your host, depending on the operations you perform on the mounted
|
||||||
filesystem.
|
filesystem.
|
||||||
|
|
||||||
To unmount the filesystem use the ``umount`` command on the mountpoint:
|
To unmount the filesystem, use the ``umount`` command on the mount point:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -506,7 +524,7 @@ To unmount the filesystem use the ``umount`` command on the mountpoint:
|
|||||||
Login and Logout
|
Login and Logout
|
||||||
----------------
|
----------------
|
||||||
|
|
||||||
The client tool prompts you to enter the logon password as soon as you
|
The client tool prompts you to enter the login password as soon as you
|
||||||
want to access the backup server. The server checks your credentials
|
want to access the backup server. The server checks your credentials
|
||||||
and responds with a ticket that is valid for two hours. The client
|
and responds with a ticket that is valid for two hours. The client
|
||||||
tool automatically stores that ticket and uses it for further requests
|
tool automatically stores that ticket and uses it for further requests
|
||||||
@ -535,7 +553,7 @@ Changing the Owner of a Backup Group
|
|||||||
By default, the owner of a backup group is the user which was used to originally
|
By default, the owner of a backup group is the user which was used to originally
|
||||||
create that backup group (or in the case of sync jobs, ``root@pam``). This
|
create that backup group (or in the case of sync jobs, ``root@pam``). This
|
||||||
means that if a user ``mike@pbs`` created a backup, another user ``john@pbs``
|
means that if a user ``mike@pbs`` created a backup, another user ``john@pbs``
|
||||||
can not be used to create backups in that same backup group. In case you want
|
can not be used to create backups in that same backup group. In case you want
|
||||||
to change the owner of a backup, you can do so with the below command, using a
|
to change the owner of a backup, you can do so with the below command, using a
|
||||||
user that has ``Datastore.Modify`` privileges on the datastore.
|
user that has ``Datastore.Modify`` privileges on the datastore.
|
||||||
|
|
||||||
@ -636,6 +654,25 @@ shows the list of existing snapshots and what actions prune would take.
|
|||||||
in the chunk-store. The chunk-store still contains the data blocks. To free
|
in the chunk-store. The chunk-store still contains the data blocks. To free
|
||||||
space you need to perform :ref:`client_garbage-collection`.
|
space you need to perform :ref:`client_garbage-collection`.
|
||||||
|
|
||||||
|
It is also possible to protect single snapshots from being pruned or deleted:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-client snapshot protected update <snapshot> true
|
||||||
|
|
||||||
|
This will set the protected flag on the snapshot and prevent pruning or manual
|
||||||
|
deletion of this snapshot untilt he flag is removed again with:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-client snapshot protected update <snapshot> false
|
||||||
|
|
||||||
|
When a group is with a protected snapshot is deleted, only the non-protected
|
||||||
|
ones are removed and the group will remain.
|
||||||
|
|
||||||
|
.. note:: This flag will not be synced when using pull or sync jobs. If you
|
||||||
|
want to protect a synced snapshot, you have to manually to this again on
|
||||||
|
the target backup server.
|
||||||
|
|
||||||
.. _client_garbage-collection:
|
.. _client_garbage-collection:
|
||||||
|
|
||||||
@ -661,7 +698,7 @@ unused data blocks are removed.
|
|||||||
(access time) property. Filesystems are mounted with the ``relatime`` option
|
(access time) property. Filesystems are mounted with the ``relatime`` option
|
||||||
by default. This results in a better performance by only updating the
|
by default. This results in a better performance by only updating the
|
||||||
``atime`` property if the last access has been at least 24 hours ago. The
|
``atime`` property if the last access has been at least 24 hours ago. The
|
||||||
downside is, that touching a chunk within these 24 hours will not always
|
downside is that touching a chunk within these 24 hours will not always
|
||||||
update its ``atime`` property.
|
update its ``atime`` property.
|
||||||
|
|
||||||
Chunks in the grace period will be logged at the end of the garbage
|
Chunks in the grace period will be logged at the end of the garbage
|
||||||
@ -685,8 +722,8 @@ unused data blocks are removed.
|
|||||||
Average chunk size: 2486565
|
Average chunk size: 2486565
|
||||||
TASK OK
|
TASK OK
|
||||||
|
|
||||||
|
Garbage collection can also be scheduled using ``promxox-backup-manager`` or
|
||||||
.. todo:: howto run garbage-collection at regular intervals (cron)
|
from the Proxmox Backup Server's web interface.
|
||||||
|
|
||||||
Benchmarking
|
Benchmarking
|
||||||
------------
|
------------
|
||||||
|
@ -1,10 +1,10 @@
|
|||||||
Backup Protocol
|
Backup Protocol
|
||||||
===============
|
===============
|
||||||
|
|
||||||
Proxmox Backup Server uses a REST based API. While the management
|
Proxmox Backup Server uses a REST-based API. While the management
|
||||||
interface use normal HTTP, the actual backup and restore interface use
|
interface uses normal HTTP, the actual backup and restore interface uses
|
||||||
HTTP/2 for improved performance. Both HTTP and HTTP/2 are well known
|
HTTP/2 for improved performance. Both HTTP and HTTP/2 are well known
|
||||||
standards, so the following section assumes that you are familiar on
|
standards, so the following section assumes that you are familiar with
|
||||||
how to use them.
|
how to use them.
|
||||||
|
|
||||||
|
|
||||||
@ -13,35 +13,35 @@ Backup Protocol API
|
|||||||
|
|
||||||
To start a new backup, the API call ``GET /api2/json/backup`` needs to
|
To start a new backup, the API call ``GET /api2/json/backup`` needs to
|
||||||
be upgraded to a HTTP/2 connection using
|
be upgraded to a HTTP/2 connection using
|
||||||
``proxmox-backup-protocol-v1`` as protocol name::
|
``proxmox-backup-protocol-v1`` as the protocol name::
|
||||||
|
|
||||||
GET /api2/json/backup HTTP/1.1
|
GET /api2/json/backup HTTP/1.1
|
||||||
UPGRADE: proxmox-backup-protocol-v1
|
UPGRADE: proxmox-backup-protocol-v1
|
||||||
|
|
||||||
The server replies with HTTP 101 Switching Protocol status code,
|
The server replies with the ``HTTP 101 Switching Protocol`` status code,
|
||||||
and you can then issue REST commands on that updated HTTP/2 connection.
|
and you can then issue REST commands on the updated HTTP/2 connection.
|
||||||
|
|
||||||
The backup protocol allows you to upload three different kind of files:
|
The backup protocol allows you to upload three different kind of files:
|
||||||
|
|
||||||
- Chunks and blobs (binary data)
|
- Chunks and blobs (binary data)
|
||||||
|
|
||||||
- Fixed Indexes (List of chunks with fixed size)
|
- Fixed indexes (List of chunks with fixed size)
|
||||||
|
|
||||||
- Dynamic Indexes (List of chunk with variable size)
|
- Dynamic indexes (List of chunks with variable size)
|
||||||
|
|
||||||
The following section gives a short introduction how to upload such
|
The following section provides a short introduction on how to upload such
|
||||||
files. Please use the `API Viewer <api-viewer/index.html>`_ for
|
files. Please use the `API Viewer <api-viewer/index.html>`_ for
|
||||||
details about available REST commands.
|
details about the available REST commands.
|
||||||
|
|
||||||
|
|
||||||
Upload Blobs
|
Upload Blobs
|
||||||
~~~~~~~~~~~~
|
~~~~~~~~~~~~
|
||||||
|
|
||||||
Uploading blobs is done using ``POST /blob``. The HTTP body contains the
|
Blobs are uploaded using ``POST /blob``. The HTTP body contains the
|
||||||
data encoded as :ref:`Data Blob <data-blob-format>`).
|
data encoded as :ref:`Data Blob <data-blob-format>`.
|
||||||
|
|
||||||
The file name needs to end with ``.blob``, and is automatically added
|
The file name must end with ``.blob``, and is automatically added
|
||||||
to the backup manifest.
|
to the backup manifest, following the call to ``POST /finish``.
|
||||||
|
|
||||||
|
|
||||||
Upload Chunks
|
Upload Chunks
|
||||||
@ -56,40 +56,41 @@ encoded as :ref:`Data Blob <data-blob-format>`).
|
|||||||
Upload Fixed Indexes
|
Upload Fixed Indexes
|
||||||
~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
Fixed indexes are use to store VM image data. The VM image is split
|
Fixed indexes are used to store VM image data. The VM image is split
|
||||||
into equally sized chunks, which are uploaded individually. The index
|
into equally sized chunks, which are uploaded individually. The index
|
||||||
file simply contains a list to chunk digests.
|
file simply contains a list of chunk digests.
|
||||||
|
|
||||||
You create a fixed index with ``POST /fixed_index``. Then upload
|
You create a fixed index with ``POST /fixed_index``. Then, upload
|
||||||
chunks with ``POST /fixed_chunk``, and append them to the index with
|
chunks with ``POST /fixed_chunk``, and append them to the index with
|
||||||
``PUT /fixed_index``. When finished, you need to close the index using
|
``PUT /fixed_index``. When finished, you need to close the index using
|
||||||
``POST /fixed_close``.
|
``POST /fixed_close``.
|
||||||
|
|
||||||
The file name needs to end with ``.fidx``, and is automatically added
|
The file name needs to end with ``.fidx``, and is automatically added
|
||||||
to the backup manifest.
|
to the backup manifest, following the call to ``POST /finish``.
|
||||||
|
|
||||||
|
|
||||||
Upload Dynamic Indexes
|
Upload Dynamic Indexes
|
||||||
~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
Dynamic indexes are use to store file archive data. The archive data
|
Dynamic indexes are used to store file archive data. The archive data
|
||||||
is split into dynamically sized chunks, which are uploaded
|
is split into dynamically sized chunks, which are uploaded
|
||||||
individually. The index file simply contains a list to chunk digests
|
individually. The index file simply contains a list of chunk digests
|
||||||
and offsets.
|
and offsets.
|
||||||
|
|
||||||
You create a dynamic sized index with ``POST /dynamic_index``. Then
|
You can create a dynamically sized index with ``POST /dynamic_index``. Then,
|
||||||
upload chunks with ``POST /dynamic_chunk``, and append them to the index with
|
upload chunks with ``POST /dynamic_chunk``, and append them to the index with
|
||||||
``PUT /dynamic_index``. When finished, you need to close the index using
|
``PUT /dynamic_index``. When finished, you need to close the index using
|
||||||
``POST /dynamic_close``.
|
``POST /dynamic_close``.
|
||||||
|
|
||||||
The file name needs to end with ``.didx``, and is automatically added
|
The filename needs to end with ``.didx``, and is automatically added
|
||||||
to the backup manifest.
|
to the backup manifest, following the call to ``POST /finish``.
|
||||||
|
|
||||||
|
|
||||||
Finish Backup
|
Finish Backup
|
||||||
~~~~~~~~~~~~~
|
~~~~~~~~~~~~~
|
||||||
|
|
||||||
Once you have uploaded all data, you need to call ``POST
|
Once you have uploaded all data, you need to call ``POST /finish``. This
|
||||||
/finish``. This commits all data and ends the backup protocol.
|
commits all data and ends the backup protocol.
|
||||||
|
|
||||||
|
|
||||||
Restore/Reader Protocol API
|
Restore/Reader Protocol API
|
||||||
@ -102,39 +103,39 @@ be upgraded to a HTTP/2 connection using
|
|||||||
GET /api2/json/reader HTTP/1.1
|
GET /api2/json/reader HTTP/1.1
|
||||||
UPGRADE: proxmox-backup-reader-protocol-v1
|
UPGRADE: proxmox-backup-reader-protocol-v1
|
||||||
|
|
||||||
The server replies with HTTP 101 Switching Protocol status code,
|
The server replies with the ``HTTP 101 Switching Protocol`` status code,
|
||||||
and you can then issue REST commands on that updated HTTP/2 connection.
|
and you can then issue REST commands on that updated HTTP/2 connection.
|
||||||
|
|
||||||
The reader protocol allows you to download three different kind of files:
|
The reader protocol allows you to download three different kinds of files:
|
||||||
|
|
||||||
- Chunks and blobs (binary data)
|
- Chunks and blobs (binary data)
|
||||||
|
|
||||||
- Fixed Indexes (List of chunks with fixed size)
|
- Fixed indexes (list of chunks with fixed size)
|
||||||
|
|
||||||
- Dynamic Indexes (List of chunk with variable size)
|
- Dynamic indexes (list of chunks with variable size)
|
||||||
|
|
||||||
The following section gives a short introduction how to download such
|
The following section provides a short introduction on how to download such
|
||||||
files. Please use the `API Viewer <api-viewer/index.html>`_ for details about
|
files. Please use the `API Viewer <api-viewer/index.html>`_ for details about
|
||||||
available REST commands.
|
the available REST commands.
|
||||||
|
|
||||||
|
|
||||||
Download Blobs
|
Download Blobs
|
||||||
~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~
|
||||||
|
|
||||||
Downloading blobs is done using ``GET /download``. The HTTP body contains the
|
Blobs are downloaded using ``GET /download``. The HTTP body contains the
|
||||||
data encoded as :ref:`Data Blob <data-blob-format>`.
|
data encoded as :ref:`Data Blob <data-blob-format>`.
|
||||||
|
|
||||||
|
|
||||||
Download Chunks
|
Download Chunks
|
||||||
~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
Downloading chunks is done using ``GET /chunk``. The HTTP body contains the
|
Chunks are downloaded using ``GET /chunk``. The HTTP body contains the
|
||||||
data encoded as :ref:`Data Blob <data-blob-format>`).
|
data encoded as :ref:`Data Blob <data-blob-format>`.
|
||||||
|
|
||||||
|
|
||||||
Download Index Files
|
Download Index Files
|
||||||
~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
Downloading index files is done using ``GET /download``. The HTTP body
|
Index files are downloaded using ``GET /download``. The HTTP body
|
||||||
contains the data encoded as :ref:`Fixed Index <fixed-index-format>`
|
contains the data encoded as :ref:`Fixed Index <fixed-index-format>`
|
||||||
or :ref:`Dynamic Index <dynamic-index-format>`.
|
or :ref:`Dynamic Index <dynamic-index-format>`.
|
||||||
|
@ -37,7 +37,7 @@ Each field can contain multiple values in the following formats:
|
|||||||
* and a combination of the above: e.g., 01,05..10,12/02
|
* and a combination of the above: e.g., 01,05..10,12/02
|
||||||
* or a `*` for every possible value: e.g., \*:00
|
* or a `*` for every possible value: e.g., \*:00
|
||||||
|
|
||||||
There are some special values that have specific meaning:
|
There are some special values that have a specific meaning:
|
||||||
|
|
||||||
================================= ==============================
|
================================= ==============================
|
||||||
Value Syntax
|
Value Syntax
|
||||||
@ -81,19 +81,19 @@ Not all features of systemd calendar events are implemented:
|
|||||||
|
|
||||||
* no Unix timestamps (e.g. `@12345`): instead use date and time to specify
|
* no Unix timestamps (e.g. `@12345`): instead use date and time to specify
|
||||||
a specific point in time
|
a specific point in time
|
||||||
* no timezone: all schedules use the set timezone on the server
|
* no timezone: all schedules use the timezone of the server
|
||||||
* no sub-second resolution
|
* no sub-second resolution
|
||||||
* no reverse day syntax (e.g. 2020-03~01)
|
* no reverse day syntax (e.g. 2020-03~01)
|
||||||
* no repetition of ranges (e.g. 1..10/2)
|
* no repetition of ranges (e.g. 1..10/2)
|
||||||
|
|
||||||
Notes on scheduling
|
Notes on Scheduling
|
||||||
-------------------
|
-------------------
|
||||||
|
|
||||||
In `Proxmox Backup`_ scheduling for most tasks is done in the
|
In `Proxmox Backup`_, scheduling for most tasks is done in the
|
||||||
`proxmox-backup-proxy`. This daemon checks all job schedules
|
`proxmox-backup-proxy`. This daemon checks all job schedules
|
||||||
if they are due every minute. This means that even if
|
every minute, to see if any are due. This means that even though
|
||||||
`calendar events` can contain seconds, it will only be checked
|
`calendar events` can contain seconds, it will only be checked
|
||||||
once a minute.
|
once per minute.
|
||||||
|
|
||||||
Also, all schedules will be checked against the timezone set
|
Also, all schedules will be checked against the timezone set
|
||||||
in the `Proxmox Backup`_ server.
|
in the `Proxmox Backup`_ server.
|
||||||
|
@ -21,3 +21,7 @@ Command Line Tools
|
|||||||
|
|
||||||
.. include:: pxar/description.rst
|
.. include:: pxar/description.rst
|
||||||
|
|
||||||
|
``proxmox-backup-debug``
|
||||||
|
~~~~~~~~
|
||||||
|
|
||||||
|
.. include:: proxmox-backup-debug/description.rst
|
||||||
|
@ -10,7 +10,7 @@ Command Syntax
|
|||||||
Catalog Shell Commands
|
Catalog Shell Commands
|
||||||
~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
Those command are available when you start an interactive restore shell:
|
The following commands are available in an interactive restore shell:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
|
@ -2,13 +2,13 @@ This file contains the access control list for the Proxmox Backup
|
|||||||
Server API.
|
Server API.
|
||||||
|
|
||||||
Each line starts with ``acl:``, followed by 4 additional values
|
Each line starts with ``acl:``, followed by 4 additional values
|
||||||
separated by collon.
|
separated by colon.
|
||||||
|
|
||||||
:propagate: Propagate permissions down the hierachrchy
|
:propagate: Propagate permissions down the hierarchy
|
||||||
|
|
||||||
:path: The object path
|
:path: The object path
|
||||||
|
|
||||||
:User/Token: List of users and token
|
:User/Token: List of users and tokens
|
||||||
|
|
||||||
:Role: List of assigned roles
|
:Role: List of assigned roles
|
||||||
|
|
||||||
|
@ -1,9 +1,9 @@
|
|||||||
The file contains a list of datastore configuration sections. Each
|
This file contains a list of datastore configuration sections. Each
|
||||||
section starts with a header ``datastore: <name>``, followed by the
|
section starts with the header ``datastore: <name>``, followed by the
|
||||||
datastore configuration options.
|
datastore configuration options.
|
||||||
|
|
||||||
::
|
::
|
||||||
|
|
||||||
datastore: <name1>
|
datastore: <name1>
|
||||||
path <path1>
|
path <path1>
|
||||||
<option1> <value1>
|
<option1> <value1>
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
Each entry starts with a header ``pool: <name>``, followed by the
|
Each entry starts with the header ``pool: <name>``, followed by the
|
||||||
media pool configuration options.
|
media pool configuration options.
|
||||||
|
|
||||||
::
|
::
|
||||||
@ -8,6 +8,6 @@ media pool configuration options.
|
|||||||
retention overwrite
|
retention overwrite
|
||||||
|
|
||||||
pool: ...
|
pool: ...
|
||||||
|
|
||||||
|
|
||||||
You can use the ``proxmox-tape pool`` command to manipulate this file.
|
You can use the ``proxmox-tape pool`` command to manipulate this file.
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
This file contains information used to access remote servers.
|
This file contains information used to access remote servers.
|
||||||
|
|
||||||
Each entry starts with a header ``remote: <name>``, followed by the
|
Each entry starts with the header ``remote: <name>``, followed by the
|
||||||
remote configuration options.
|
remote configuration options.
|
||||||
|
|
||||||
::
|
::
|
||||||
@ -11,7 +11,7 @@ remote configuration options.
|
|||||||
...
|
...
|
||||||
|
|
||||||
remote: ...
|
remote: ...
|
||||||
|
|
||||||
|
|
||||||
You can use the ``proxmox-backup-manager remote`` command to manipulate
|
You can use the ``proxmox-backup-manager remote`` command to manipulate
|
||||||
this file.
|
this file.
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
Each entry starts with a header ``sync: <name>``, followed by the
|
Each entry starts with the header ``sync: <name>``, followed by the
|
||||||
job configuration options.
|
job configuration options.
|
||||||
|
|
||||||
::
|
::
|
||||||
@ -9,7 +9,7 @@ job configuration options.
|
|||||||
remote lina
|
remote lina
|
||||||
|
|
||||||
sync: ...
|
sync: ...
|
||||||
|
|
||||||
|
|
||||||
You can use the ``proxmox-backup-manager sync-job`` command to manipulate
|
You can use the ``proxmox-backup-manager sync-job`` command to manipulate
|
||||||
this file.
|
this file.
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
Each entry starts with a header ``backup: <name>``, followed by the
|
Each entry starts with the header ``backup: <name>``, followed by the
|
||||||
job configuration options.
|
job configuration options.
|
||||||
|
|
||||||
::
|
::
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
Each LTO drive configuration section starts with a header ``lto: <name>``,
|
Each LTO drive configuration section starts with the header ``lto: <name>``,
|
||||||
followed by the drive configuration options.
|
followed by the drive configuration options.
|
||||||
|
|
||||||
Tape changer configurations starts with ``changer: <name>``,
|
Tape changer configurations start with the header ``changer: <name>``,
|
||||||
followed by the changer configuration options.
|
followed by the changer configuration options.
|
||||||
|
|
||||||
::
|
::
|
||||||
@ -18,5 +18,5 @@ followed by the changer configuration options.
|
|||||||
You can use the ``proxmox-tape drive`` and ``proxmox-tape changer``
|
You can use the ``proxmox-tape drive`` and ``proxmox-tape changer``
|
||||||
commands to manipulate this file.
|
commands to manipulate this file.
|
||||||
|
|
||||||
.. NOTE:: The ``virtual:`` drive type is experimental and onyl used
|
.. NOTE:: The ``virtual:`` drive type is experimental and should only be used
|
||||||
for debugging.
|
for debugging.
|
||||||
|
@ -1,9 +1,9 @@
|
|||||||
This file contains the list of API users and API tokens.
|
This file contains the list of API users and API tokens.
|
||||||
|
|
||||||
Each user configuration section starts with a header ``user: <name>``,
|
Each user configuration section starts with the header ``user: <name>``,
|
||||||
followed by the user configuration options.
|
followed by the user configuration options.
|
||||||
|
|
||||||
API token configuration starts with a header ``token:
|
API token configuration starts with the header ``token:
|
||||||
<userid!token_name>``, followed by the token configuration. The data
|
<userid!token_name>``, followed by the token configuration. The data
|
||||||
used to authenticate tokens is stored in a separate file
|
used to authenticate tokens is stored in a separate file
|
||||||
(``token.shadow``).
|
(``token.shadow``).
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
Each entry starts with a header ``verification: <name>``, followed by the
|
Each entry starts with the header ``verification: <name>``, followed by the
|
||||||
job configuration options.
|
job configuration options.
|
||||||
|
|
||||||
::
|
::
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
Configuration Files
|
Configuration Files
|
||||||
===================
|
===================
|
||||||
|
|
||||||
All Proxmox Backup Server configuration files resides inside directory
|
All Proxmox Backup Server configuration files reside in the directory
|
||||||
``/etc/proxmox-backup/``.
|
``/etc/proxmox-backup/``.
|
||||||
|
|
||||||
|
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
.. Epilog (included at top of each file)
|
.. Epilog (included at top of each file)
|
||||||
|
|
||||||
We use this file to define external links and common replacement
|
We use this file to define external links and common replacement
|
||||||
patterns.
|
patterns.
|
||||||
|
|
||||||
@ -13,7 +13,6 @@
|
|||||||
.. _Proxmox: https://www.proxmox.com
|
.. _Proxmox: https://www.proxmox.com
|
||||||
.. _Proxmox Community Forum: https://forum.proxmox.com
|
.. _Proxmox Community Forum: https://forum.proxmox.com
|
||||||
.. _Proxmox Virtual Environment: https://www.proxmox.com/proxmox-ve
|
.. _Proxmox Virtual Environment: https://www.proxmox.com/proxmox-ve
|
||||||
.. FIXME
|
|
||||||
.. _Proxmox Backup: https://pbs.proxmox.com/wiki/index.php/Main_Page
|
.. _Proxmox Backup: https://pbs.proxmox.com/wiki/index.php/Main_Page
|
||||||
.. _PBS Development List: https://lists.proxmox.com/cgi-bin/mailman/listinfo/pbs-devel
|
.. _PBS Development List: https://lists.proxmox.com/cgi-bin/mailman/listinfo/pbs-devel
|
||||||
.. _reStructuredText: https://www.sphinx-doc.org/en/master/usage/restructuredtext/index.html
|
.. _reStructuredText: https://www.sphinx-doc.org/en/master/usage/restructuredtext/index.html
|
||||||
@ -23,6 +22,7 @@
|
|||||||
.. _Virtual machine: https://en.wikipedia.org/wiki/Virtual_machine
|
.. _Virtual machine: https://en.wikipedia.org/wiki/Virtual_machine
|
||||||
.. _APT: http://en.wikipedia.org/wiki/Advanced_Packaging_Tool
|
.. _APT: http://en.wikipedia.org/wiki/Advanced_Packaging_Tool
|
||||||
.. _QEMU: https://www.qemu.org/
|
.. _QEMU: https://www.qemu.org/
|
||||||
|
.. _LXC: https://linuxcontainers.org/lxc/introduction/
|
||||||
|
|
||||||
.. _Client-server model: https://en.wikipedia.org/wiki/Client-server_model
|
.. _Client-server model: https://en.wikipedia.org/wiki/Client-server_model
|
||||||
.. _AE: https://en.wikipedia.org/wiki/Authenticated_encryption
|
.. _AE: https://en.wikipedia.org/wiki/Authenticated_encryption
|
||||||
|
18
docs/faq.rst
18
docs/faq.rst
@ -24,11 +24,13 @@ future plans to support 32-bit processors.
|
|||||||
How long will my Proxmox Backup Server version be supported?
|
How long will my Proxmox Backup Server version be supported?
|
||||||
------------------------------------------------------------
|
------------------------------------------------------------
|
||||||
|
|
||||||
+-----------------------+--------------------+---------------+------------+--------------------+
|
+-----------------------+----------------------+---------------+------------+--------------------+
|
||||||
|Proxmox Backup Version | Debian Version | First Release | Debian EOL | Proxmox Backup EOL |
|
|Proxmox Backup Version | Debian Version | First Release | Debian EOL | Proxmox Backup EOL |
|
||||||
+=======================+====================+===============+============+====================+
|
+=======================+======================+===============+============+====================+
|
||||||
|Proxmox Backup 1.x | Debian 10 (Buster) | 2020-11 | tba | tba |
|
|Proxmox Backup 2.x | Debian 11 (Bullseye) | 2021-07 | tba | tba |
|
||||||
+-----------------------+--------------------+---------------+------------+--------------------+
|
+-----------------------+----------------------+---------------+------------+--------------------+
|
||||||
|
|Proxmox Backup 1.x | Debian 10 (Buster) | 2020-11 | ~Q2/2022 | Q2-Q3/2022 |
|
||||||
|
+-----------------------+----------------------+---------------+------------+--------------------+
|
||||||
|
|
||||||
|
|
||||||
Can I copy or synchronize my datastore to another location?
|
Can I copy or synchronize my datastore to another location?
|
||||||
@ -67,6 +69,6 @@ be able to read the data.
|
|||||||
Is the backup incremental/deduplicated?
|
Is the backup incremental/deduplicated?
|
||||||
---------------------------------------
|
---------------------------------------
|
||||||
|
|
||||||
With Proxmox Backup Server, backups are sent incremental and data is
|
With Proxmox Backup Server, backups are sent incrementally to the server, and
|
||||||
deduplicated on the server.
|
data is then deduplicated on the server. This minimizes both the storage
|
||||||
This minimizes both the storage consumed and the network impact.
|
consumed and the impact on the network.
|
||||||
|
@ -14,7 +14,8 @@ Proxmox File Archive Format (``.pxar``)
|
|||||||
Data Blob Format (``.blob``)
|
Data Blob Format (``.blob``)
|
||||||
----------------------------
|
----------------------------
|
||||||
|
|
||||||
The data blob format is used to store small binary data. The magic number decides the exact format:
|
The data blob format is used to store small binary data. The magic number
|
||||||
|
decides the exact format:
|
||||||
|
|
||||||
.. list-table::
|
.. list-table::
|
||||||
:widths: auto
|
:widths: auto
|
||||||
@ -32,7 +33,8 @@ The data blob format is used to store small binary data. The magic number decide
|
|||||||
- encrypted
|
- encrypted
|
||||||
- compressed
|
- compressed
|
||||||
|
|
||||||
Compression algorithm is ``zstd``. Encryption cipher is ``AES_256_GCM``.
|
The compression algorithm used is ``zstd``. The encryption cipher is
|
||||||
|
``AES_256_GCM``.
|
||||||
|
|
||||||
Unencrypted blobs use the following format:
|
Unencrypted blobs use the following format:
|
||||||
|
|
||||||
@ -43,15 +45,15 @@ Unencrypted blobs use the following format:
|
|||||||
* - ``CRC32: [u8; 4]``
|
* - ``CRC32: [u8; 4]``
|
||||||
* - ``Data: (max 16MiB)``
|
* - ``Data: (max 16MiB)``
|
||||||
|
|
||||||
Encrypted blobs additionally contains a 16 byte IV, followed by a 16
|
Encrypted blobs additionally contain a 16 byte initialization vector (IV),
|
||||||
byte Authenticated Encyryption (AE) tag, followed by the encrypted
|
followed by a 16 byte authenticated encryption (AE) tag, followed by the
|
||||||
data:
|
encrypted data:
|
||||||
|
|
||||||
.. list-table::
|
.. list-table::
|
||||||
|
|
||||||
* - ``MAGIC: [u8; 8]``
|
* - ``MAGIC: [u8; 8]``
|
||||||
* - ``CRC32: [u8; 4]``
|
* - ``CRC32: [u8; 4]``
|
||||||
* - ``ÌV: [u8; 16]``
|
* - ``IV: [u8; 16]``
|
||||||
* - ``TAG: [u8; 16]``
|
* - ``TAG: [u8; 16]``
|
||||||
* - ``Data: (max 16MiB)``
|
* - ``Data: (max 16MiB)``
|
||||||
|
|
||||||
@ -72,19 +74,19 @@ All numbers are stored as little-endian.
|
|||||||
* - ``ctime: i64``,
|
* - ``ctime: i64``,
|
||||||
- Creation Time (epoch)
|
- Creation Time (epoch)
|
||||||
* - ``index_csum: [u8; 32]``,
|
* - ``index_csum: [u8; 32]``,
|
||||||
- Sha256 over the index (without header) ``SHA256(digest1||digest2||...)``
|
- SHA-256 over the index (without header) ``SHA256(digest1||digest2||...)``
|
||||||
* - ``size: u64``,
|
* - ``size: u64``,
|
||||||
- Image size
|
- Image size
|
||||||
* - ``chunk_size: u64``,
|
* - ``chunk_size: u64``,
|
||||||
- Chunk size
|
- Chunk size
|
||||||
* - ``reserved: [u8; 4016]``,
|
* - ``reserved: [u8; 4016]``,
|
||||||
- overall header size is one page (4096 bytes)
|
- Overall header size is one page (4096 bytes)
|
||||||
* - ``digest1: [u8; 32]``
|
* - ``digest1: [u8; 32]``
|
||||||
- first chunk digest
|
- First chunk digest
|
||||||
* - ``digest2: [u8; 32]``
|
* - ``digest2: [u8; 32]``
|
||||||
- next chunk
|
- Second chunk digest
|
||||||
* - ...
|
* - ...
|
||||||
- next chunk ...
|
- Next chunk digest ...
|
||||||
|
|
||||||
|
|
||||||
.. _dynamic-index-format:
|
.. _dynamic-index-format:
|
||||||
@ -103,16 +105,16 @@ All numbers are stored as little-endian.
|
|||||||
* - ``ctime: i64``,
|
* - ``ctime: i64``,
|
||||||
- Creation Time (epoch)
|
- Creation Time (epoch)
|
||||||
* - ``index_csum: [u8; 32]``,
|
* - ``index_csum: [u8; 32]``,
|
||||||
- Sha256 over the index (without header) ``SHA256(offset1||digest1||offset2||digest2||...)``
|
- SHA-256 over the index (without header) ``SHA256(offset1||digest1||offset2||digest2||...)``
|
||||||
* - ``reserved: [u8; 4032]``,
|
* - ``reserved: [u8; 4032]``,
|
||||||
- Overall header size is one page (4096 bytes)
|
- Overall header size is one page (4096 bytes)
|
||||||
* - ``offset1: u64``
|
* - ``offset1: u64``
|
||||||
- End of first chunk
|
- End of first chunk
|
||||||
* - ``digest1: [u8; 32]``
|
* - ``digest1: [u8; 32]``
|
||||||
- first chunk digest
|
- First chunk digest
|
||||||
* - ``offset2: u64``
|
* - ``offset2: u64``
|
||||||
- End of second chunk
|
- End of second chunk
|
||||||
* - ``digest2: [u8; 32]``
|
* - ``digest2: [u8; 32]``
|
||||||
- second chunk digest
|
- Second chunk digest
|
||||||
* - ...
|
* - ...
|
||||||
- next chunk offset/digest
|
- Next chunk offset/digest
|
||||||
|
@ -11,7 +11,7 @@ Glossary
|
|||||||
`Container`_
|
`Container`_
|
||||||
|
|
||||||
A container is an isolated user space. Programs run directly on
|
A container is an isolated user space. Programs run directly on
|
||||||
the host's kernel, but with limited access to the host resources.
|
the host's kernel, but with limited access to the host's resources.
|
||||||
|
|
||||||
Datastore
|
Datastore
|
||||||
|
|
||||||
@ -23,19 +23,19 @@ Glossary
|
|||||||
Rust is a new, fast and memory-efficient system programming
|
Rust is a new, fast and memory-efficient system programming
|
||||||
language. It has no runtime or garbage collector. Rust’s rich type
|
language. It has no runtime or garbage collector. Rust’s rich type
|
||||||
system and ownership model guarantee memory-safety and
|
system and ownership model guarantee memory-safety and
|
||||||
thread-safety. I can eliminate many classes of bugs
|
thread-safety. This can eliminate many classes of bugs
|
||||||
at compile-time.
|
at compile-time.
|
||||||
|
|
||||||
`Sphinx`_
|
`Sphinx`_
|
||||||
|
|
||||||
Is a tool that makes it easy to create intelligent and
|
Is a tool that makes it easy to create intelligent and nicely formatted
|
||||||
beautiful documentation. It was originally created for the
|
documentation. It was originally created for the documentation of the
|
||||||
documentation of the Python programming language. It has excellent facilities for the
|
Python programming language. It has excellent facilities for the
|
||||||
documentation of software projects in a range of languages.
|
documentation of software projects in a range of languages.
|
||||||
|
|
||||||
`reStructuredText`_
|
`reStructuredText`_
|
||||||
|
|
||||||
Is an easy-to-read, what-you-see-is-what-you-get plaintext
|
Is an easy-to-read, what-you-see-is-what-you-get, plaintext
|
||||||
markup syntax and parser system.
|
markup syntax and parser system.
|
||||||
|
|
||||||
`FUSE`
|
`FUSE`
|
||||||
|
54
docs/gui.rst
54
docs/gui.rst
@ -8,8 +8,9 @@ tools. The web interface also provides a built-in console, so if you prefer the
|
|||||||
command line or need some extra control, you have this option.
|
command line or need some extra control, you have this option.
|
||||||
|
|
||||||
The web interface can be accessed via https://youripaddress:8007. The default
|
The web interface can be accessed via https://youripaddress:8007. The default
|
||||||
login is `root`, and the password is the one specified during the installation
|
login is `root`, and the password is either the one specified during the
|
||||||
process.
|
installation process or the password of the root user, in case of installation
|
||||||
|
on top of Debian.
|
||||||
|
|
||||||
|
|
||||||
Features
|
Features
|
||||||
@ -48,12 +49,13 @@ GUI Overview
|
|||||||
|
|
||||||
The Proxmox Backup Server web interface consists of 3 main sections:
|
The Proxmox Backup Server web interface consists of 3 main sections:
|
||||||
|
|
||||||
* **Header**: At the top. This shows version information, and contains buttons to view
|
* **Header**: At the top. This shows version information and contains buttons to
|
||||||
documentation, monitor running tasks, set the language and logout.
|
view documentation, monitor running tasks, set the language, configure various
|
||||||
* **Sidebar**: On the left. This contains the configuration options for
|
display settings, and logout.
|
||||||
|
* **Sidebar**: On the left. This contains the administration options for
|
||||||
the server.
|
the server.
|
||||||
* **Configuration Panel**: In the center. This contains the control interface for the
|
* **Configuration Panel**: In the center. This contains the respective control
|
||||||
configuration options in the *Sidebar*.
|
interfaces for the administration options in the *Sidebar*.
|
||||||
|
|
||||||
|
|
||||||
Sidebar
|
Sidebar
|
||||||
@ -74,12 +76,14 @@ previous and currently running tasks, and subscription information.
|
|||||||
Configuration
|
Configuration
|
||||||
^^^^^^^^^^^^^
|
^^^^^^^^^^^^^
|
||||||
|
|
||||||
The Configuration section contains some system configuration options, such as
|
The Configuration section contains some system options, such as time, network,
|
||||||
time and network configuration. It also contains the following subsections:
|
WebAuthn, and HTTP proxy configuration. It also contains the following
|
||||||
|
subsections:
|
||||||
|
|
||||||
* **Access Control**: Add and manage users, API tokens, and the permissions
|
* **Access Control**: Add and manage users, API tokens, and the permissions
|
||||||
associated with these items
|
associated with these items
|
||||||
* **Remotes**: Add, edit and remove remotes (see :term:`Remote`)
|
* **Remotes**: Add, edit and remove remotes (see :term:`Remote`)
|
||||||
|
* **Certificates**: Manage ACME accounts and create SSL certificates.
|
||||||
* **Subscription**: Upload a subscription key, view subscription status and
|
* **Subscription**: Upload a subscription key, view subscription status and
|
||||||
access a text-based system report.
|
access a text-based system report.
|
||||||
|
|
||||||
@ -98,6 +102,7 @@ tasks and information. These are:
|
|||||||
resource usage statistics
|
resource usage statistics
|
||||||
* **Services**: Manage and monitor system services
|
* **Services**: Manage and monitor system services
|
||||||
* **Updates**: An interface for upgrading packages
|
* **Updates**: An interface for upgrading packages
|
||||||
|
* **Repositories**: An interface for configuring APT repositories
|
||||||
* **Syslog**: View log messages from the server
|
* **Syslog**: View log messages from the server
|
||||||
* **Tasks**: Task history with multiple filter options
|
* **Tasks**: Task history with multiple filter options
|
||||||
|
|
||||||
@ -110,7 +115,7 @@ The administration menu item also contains a disk management subsection:
|
|||||||
* **Disks**: View information on available disks
|
* **Disks**: View information on available disks
|
||||||
|
|
||||||
* **Directory**: Create and view information on *ext4* and *xfs* disks
|
* **Directory**: Create and view information on *ext4* and *xfs* disks
|
||||||
* **ZFS**: Create and view information on *ZFS* disks
|
* **ZFS**: Create and view information on *ZFS* disks
|
||||||
|
|
||||||
Tape Backup
|
Tape Backup
|
||||||
^^^^^^^^^^^
|
^^^^^^^^^^^
|
||||||
@ -119,11 +124,20 @@ Tape Backup
|
|||||||
:align: right
|
:align: right
|
||||||
:alt: Tape Backup: Tape changer overview
|
:alt: Tape Backup: Tape changer overview
|
||||||
|
|
||||||
The `Tape Backup`_ section contains a top panel, managing tape media sets,
|
The `Tape Backup`_ section contains a top panel, with options for managing tape
|
||||||
inventories, drives, changers and the tape backup jobs itself.
|
media sets, inventories, drives, changers, encryption keys, and the tape backup
|
||||||
|
jobs itself. The tabs are as follows:
|
||||||
|
|
||||||
It also contains a subsection per standalone drive and per changer, with a
|
* **Content**: Information on the contents of the tape backup
|
||||||
status and management view for those devices.
|
* **Inventory**: Manage the tapes attached to the system
|
||||||
|
* **Changers**: Manage tape loading devices
|
||||||
|
* **Drives**: Manage drives used for reading and writing to tapes
|
||||||
|
* **Media Pools**: Manage logical pools of tapes
|
||||||
|
* **Encryption Keys**: Manage tape backup encryption keys
|
||||||
|
* **Backup Jobs**: Manage tape backup jobs
|
||||||
|
|
||||||
|
The section also contains a subsection per standalone drive and per changer,
|
||||||
|
with a status and management view for those devices.
|
||||||
|
|
||||||
Datastore
|
Datastore
|
||||||
^^^^^^^^^
|
^^^^^^^^^
|
||||||
@ -133,9 +147,9 @@ Datastore
|
|||||||
:alt: Datastore Configuration
|
:alt: Datastore Configuration
|
||||||
|
|
||||||
The Datastore section contains interfaces for creating and managing
|
The Datastore section contains interfaces for creating and managing
|
||||||
datastores. It contains a button to create a new datastore on the server, as
|
datastores. It also contains a button for creating a new datastore on the
|
||||||
well as a subsection for each datastore on the system, in which you can use the
|
server, as well as a subsection for each datastore on the system, in which you
|
||||||
top panel to view:
|
can use the top panel to view:
|
||||||
|
|
||||||
* **Summary**: Access a range of datastore usage statistics
|
* **Summary**: Access a range of datastore usage statistics
|
||||||
* **Content**: Information on the datastore's backup groups and their respective
|
* **Content**: Information on the datastore's backup groups and their respective
|
||||||
@ -144,5 +158,7 @@ top panel to view:
|
|||||||
collection <client_garbage-collection>` operations, and run garbage collection
|
collection <client_garbage-collection>` operations, and run garbage collection
|
||||||
manually
|
manually
|
||||||
* **Sync Jobs**: Create, manage and run :ref:`syncjobs` from remote servers
|
* **Sync Jobs**: Create, manage and run :ref:`syncjobs` from remote servers
|
||||||
* **Verify Jobs**: Create, manage and run :ref:`maintenance_verification` jobs on the
|
* **Verify Jobs**: Create, manage and run :ref:`maintenance_verification` jobs
|
||||||
datastore
|
on the datastore
|
||||||
|
* **Options**: Configure notification and verification settings
|
||||||
|
* **Permissions**: Manage permissions on the datastore
|
||||||
|
Binary file not shown.
Before Width: | Height: | Size: 21 KiB After Width: | Height: | Size: 24 KiB |
BIN
docs/images/screenshots/pbs-gui-traffic-control-add.png
Normal file
BIN
docs/images/screenshots/pbs-gui-traffic-control-add.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 32 KiB |
@ -19,24 +19,24 @@ for various management tasks such as disk management.
|
|||||||
`Proxmox Backup`_ without the server part.
|
`Proxmox Backup`_ without the server part.
|
||||||
|
|
||||||
The disk image (ISO file) provided by Proxmox includes a complete Debian system
|
The disk image (ISO file) provided by Proxmox includes a complete Debian system
|
||||||
("buster" for version 1.x) as well as all necessary packages for the `Proxmox Backup`_ server.
|
as well as all necessary packages for the `Proxmox Backup`_ Server.
|
||||||
|
|
||||||
The installer will guide you through the setup process and allow
|
The installer will guide you through the setup process and allow
|
||||||
you to partition the local disk(s), apply basic system configurations
|
you to partition the local disk(s), apply basic system configuration
|
||||||
(e.g. timezone, language, network), and install all required packages.
|
(for example timezone, language, network), and install all required packages.
|
||||||
The provided ISO will get you started in just a few minutes, and is the
|
The provided ISO will get you started in just a few minutes, and is the
|
||||||
recommended method for new and existing users.
|
recommended method for new and existing users.
|
||||||
|
|
||||||
Alternatively, `Proxmox Backup`_ server can be installed on top of an
|
Alternatively, `Proxmox Backup`_ Server can be installed on top of an
|
||||||
existing Debian system.
|
existing Debian system.
|
||||||
|
|
||||||
Install `Proxmox Backup`_ with the Installer
|
Install `Proxmox Backup`_ Server using the Installer
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
Download the ISO from |DOWNLOADS|.
|
Download the ISO from |DOWNLOADS|.
|
||||||
It includes the following:
|
It includes the following:
|
||||||
|
|
||||||
* The `Proxmox Backup`_ server installer, which partitions the local
|
* The `Proxmox Backup`_ Server installer, which partitions the local
|
||||||
disk(s) with ext4, xfs or ZFS, and installs the operating system
|
disk(s) with ext4, xfs or ZFS, and installs the operating system
|
||||||
|
|
||||||
* Complete operating system (Debian Linux, 64-bit)
|
* Complete operating system (Debian Linux, 64-bit)
|
||||||
@ -63,7 +63,7 @@ standard Debian installation. After configuring the
|
|||||||
# apt-get update
|
# apt-get update
|
||||||
# apt-get install proxmox-backup-server
|
# apt-get install proxmox-backup-server
|
||||||
|
|
||||||
The commands above keep the current (Debian) kernel and install a minimal
|
The above commands keep the current (Debian) kernel and install a minimal
|
||||||
set of required packages.
|
set of required packages.
|
||||||
|
|
||||||
If you want to install the same set of packages as the installer
|
If you want to install the same set of packages as the installer
|
||||||
|
@ -4,15 +4,15 @@ Introduction
|
|||||||
What is Proxmox Backup Server?
|
What is Proxmox Backup Server?
|
||||||
------------------------------
|
------------------------------
|
||||||
|
|
||||||
Proxmox Backup Server is an enterprise-class, client-server backup software
|
Proxmox Backup Server is an enterprise-class, client-server backup solution that
|
||||||
package that backs up :term:`virtual machine`\ s, :term:`container`\ s, and
|
is capable of backing up :term:`virtual machine`\ s, :term:`container`\ s, and
|
||||||
physical hosts. It is specially optimized for the `Proxmox Virtual Environment`_
|
physical hosts. It is specially optimized for the `Proxmox Virtual Environment`_
|
||||||
platform and allows you to back up your data securely, even between remote
|
platform and allows you to back up your data securely, even between remote
|
||||||
sites, providing easy management with a web-based user interface.
|
sites, providing easy management through a web-based user interface.
|
||||||
|
|
||||||
It supports deduplication, compression, and authenticated
|
It supports deduplication, compression, and authenticated
|
||||||
encryption (AE_). Using :term:`Rust` as the implementation language guarantees high
|
encryption (AE_). Using :term:`Rust` as the implementation language guarantees
|
||||||
performance, low resource usage, and a safe, high-quality codebase.
|
high performance, low resource usage, and a safe, high-quality codebase.
|
||||||
|
|
||||||
Proxmox Backup uses state of the art cryptography for both client-server
|
Proxmox Backup uses state of the art cryptography for both client-server
|
||||||
communication and backup content :ref:`encryption <client_encryption>`. All
|
communication and backup content :ref:`encryption <client_encryption>`. All
|
||||||
@ -28,22 +28,23 @@ Proxmox Backup Server uses a `client-server model`_. The server stores the
|
|||||||
backup data and provides an API to create and manage datastores. With the
|
backup data and provides an API to create and manage datastores. With the
|
||||||
API, it's also possible to manage disks and other server-side resources.
|
API, it's also possible to manage disks and other server-side resources.
|
||||||
|
|
||||||
The backup client uses this API to access the backed up data. With the command
|
The backup client uses this API to access the backed up data. You can use the
|
||||||
line tool ``proxmox-backup-client`` you can create backups and restore data.
|
``proxmox-backup-client`` command line tool to create and restore file backups.
|
||||||
For QEMU_ with `Proxmox Virtual Environment`_ we deliver an integrated client.
|
For QEMU_ and LXC_ within `Proxmox Virtual Environment`_, we deliver an
|
||||||
|
integrated client.
|
||||||
|
|
||||||
A single backup is allowed to contain several archives. For example, when you
|
A single backup is allowed to contain several archives. For example, when you
|
||||||
backup a :term:`virtual machine`, each disk is stored as a separate archive
|
backup a :term:`virtual machine`, each disk is stored as a separate archive
|
||||||
inside that backup. The VM configuration itself is stored as an extra file.
|
inside that backup. The VM configuration itself is stored as an extra file.
|
||||||
This way, it's easy to access and restore only important parts of the backup,
|
This way, it's easy to access and restore only the important parts of the
|
||||||
without the need to scan the whole backup.
|
backup, without the need to scan the whole backup.
|
||||||
|
|
||||||
|
|
||||||
Main Features
|
Main Features
|
||||||
-------------
|
-------------
|
||||||
|
|
||||||
:Support for Proxmox VE: The `Proxmox Virtual Environment`_ is fully
|
:Support for Proxmox VE: The `Proxmox Virtual Environment`_ is fully
|
||||||
supported and you can easily backup :term:`virtual machine`\ s and
|
supported, and you can easily backup :term:`virtual machine`\ s and
|
||||||
:term:`container`\ s.
|
:term:`container`\ s.
|
||||||
|
|
||||||
:Performance: The whole software stack is written in :term:`Rust`,
|
:Performance: The whole software stack is written in :term:`Rust`,
|
||||||
@ -70,6 +71,10 @@ Main Features
|
|||||||
modern hardware. In addition to client-side encryption, all data is
|
modern hardware. In addition to client-side encryption, all data is
|
||||||
transferred via a secure TLS connection.
|
transferred via a secure TLS connection.
|
||||||
|
|
||||||
|
:Tape backup: For long-term archiving of data, Proxmox Backup Server also
|
||||||
|
provides extensive support for backing up to tape and managing tape
|
||||||
|
libraries.
|
||||||
|
|
||||||
:Web interface: Manage the Proxmox Backup Server with the integrated, web-based
|
:Web interface: Manage the Proxmox Backup Server with the integrated, web-based
|
||||||
user interface.
|
user interface.
|
||||||
|
|
||||||
@ -80,7 +85,7 @@ Main Features
|
|||||||
backup-clients.
|
backup-clients.
|
||||||
|
|
||||||
:Enterprise Support: Proxmox Server Solutions GmbH offers enterprise support in
|
:Enterprise Support: Proxmox Server Solutions GmbH offers enterprise support in
|
||||||
form of `Proxmox Backup Server Subscription Plans
|
the form of `Proxmox Backup Server Subscription Plans
|
||||||
<https://www.proxmox.com/en/proxmox-backup-server/pricing>`_. Users at every
|
<https://www.proxmox.com/en/proxmox-backup-server/pricing>`_. Users at every
|
||||||
subscription level get access to the Proxmox Backup :ref:`Enterprise
|
subscription level get access to the Proxmox Backup :ref:`Enterprise
|
||||||
Repository <sysadmin_package_repos_enterprise>`. In addition, with a Basic,
|
Repository <sysadmin_package_repos_enterprise>`. In addition, with a Basic,
|
||||||
@ -173,7 +178,7 @@ Bug Tracker
|
|||||||
~~~~~~~~~~~
|
~~~~~~~~~~~
|
||||||
|
|
||||||
Proxmox runs a public bug tracker at `<https://bugzilla.proxmox.com>`_. If an
|
Proxmox runs a public bug tracker at `<https://bugzilla.proxmox.com>`_. If an
|
||||||
issue appears, file your report there. An issue can be a bug as well as a
|
issue appears, file your report there. An issue can be a bug, as well as a
|
||||||
request for a new feature or enhancement. The bug tracker helps to keep track
|
request for a new feature or enhancement. The bug tracker helps to keep track
|
||||||
of the issue and will send a notification once it has been solved.
|
of the issue and will send a notification once it has been solved.
|
||||||
|
|
||||||
@ -224,5 +229,6 @@ requirements.
|
|||||||
|
|
||||||
In July 2020, we released the first beta version of Proxmox Backup
|
In July 2020, we released the first beta version of Proxmox Backup
|
||||||
Server, followed by the first stable version in November 2020. With support for
|
Server, followed by the first stable version in November 2020. With support for
|
||||||
incremental, fully deduplicated backups, Proxmox Backup significantly reduces
|
encryption and incremental, fully deduplicated backups, Proxmox Backup offers a
|
||||||
network load and saves valuable storage space.
|
secure environment, which significantly reduces network load and saves valuable
|
||||||
|
storage space.
|
||||||
|
@ -4,17 +4,17 @@
|
|||||||
ZFS on Linux
|
ZFS on Linux
|
||||||
------------
|
------------
|
||||||
|
|
||||||
ZFS is a combined file system and logical volume manager designed by
|
ZFS is a combined file system and logical volume manager, designed by
|
||||||
Sun Microsystems. There is no need to manually compile ZFS modules - all
|
Sun Microsystems. There is no need to manually compile ZFS modules - all
|
||||||
packages are included.
|
packages are included.
|
||||||
|
|
||||||
By using ZFS, it's possible to achieve maximum enterprise features with
|
By using ZFS, it's possible to achieve maximum enterprise features with
|
||||||
low budget hardware, but also high performance systems by leveraging
|
low budget hardware, and also high performance systems by leveraging
|
||||||
SSD caching or even SSD only setups. ZFS can replace cost intense
|
SSD caching or even SSD only setups. ZFS can replace expensive
|
||||||
hardware raid cards by moderate CPU and memory load combined with easy
|
hardware raid cards with moderate CPU and memory load, combined with easy
|
||||||
management.
|
management.
|
||||||
|
|
||||||
General ZFS advantages
|
General advantages of ZFS:
|
||||||
|
|
||||||
* Easy configuration and management with GUI and CLI.
|
* Easy configuration and management with GUI and CLI.
|
||||||
* Reliable
|
* Reliable
|
||||||
@ -34,18 +34,18 @@ General ZFS advantages
|
|||||||
Hardware
|
Hardware
|
||||||
~~~~~~~~~
|
~~~~~~~~~
|
||||||
|
|
||||||
ZFS depends heavily on memory, so you need at least 8GB to start. In
|
ZFS depends heavily on memory, so it's recommended to have at least 8GB to
|
||||||
practice, use as much you can get for your hardware/budget. To prevent
|
start. In practice, use as much you can get for your hardware/budget. To prevent
|
||||||
data corruption, we recommend the use of high quality ECC RAM.
|
data corruption, we recommend the use of high quality ECC RAM.
|
||||||
|
|
||||||
If you use a dedicated cache and/or log disk, you should use an
|
If you use a dedicated cache and/or log disk, you should use an
|
||||||
enterprise class SSD (e.g. Intel SSD DC S3700 Series). This can
|
enterprise class SSD (for example, Intel SSD DC S3700 Series). This can
|
||||||
increase the overall performance significantly.
|
increase the overall performance significantly.
|
||||||
|
|
||||||
IMPORTANT: Do not use ZFS on top of hardware controller which has its
|
IMPORTANT: Do not use ZFS on top of a hardware controller which has its
|
||||||
own cache management. ZFS needs to directly communicate with disks. An
|
own cache management. ZFS needs to directly communicate with disks. An
|
||||||
HBA adapter is the way to go, or something like LSI controller flashed
|
HBA adapter or something like an LSI controller flashed in ``IT`` mode is
|
||||||
in ``IT`` mode.
|
recommended.
|
||||||
|
|
||||||
|
|
||||||
ZFS Administration
|
ZFS Administration
|
||||||
@ -53,7 +53,7 @@ ZFS Administration
|
|||||||
|
|
||||||
This section gives you some usage examples for common tasks. ZFS
|
This section gives you some usage examples for common tasks. ZFS
|
||||||
itself is really powerful and provides many options. The main commands
|
itself is really powerful and provides many options. The main commands
|
||||||
to manage ZFS are `zfs` and `zpool`. Both commands come with great
|
to manage ZFS are `zfs` and `zpool`. Both commands come with extensive
|
||||||
manual pages, which can be read with:
|
manual pages, which can be read with:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
@ -123,7 +123,7 @@ Create a new pool with cache (L2ARC)
|
|||||||
It is possible to use a dedicated cache drive partition to increase
|
It is possible to use a dedicated cache drive partition to increase
|
||||||
the performance (use SSD).
|
the performance (use SSD).
|
||||||
|
|
||||||
As `<device>` it is possible to use more devices, like it's shown in
|
For `<device>`, you can use multiple devices, as is shown in
|
||||||
"Create a new pool with RAID*".
|
"Create a new pool with RAID*".
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
@ -136,7 +136,7 @@ Create a new pool with log (ZIL)
|
|||||||
It is possible to use a dedicated cache drive partition to increase
|
It is possible to use a dedicated cache drive partition to increase
|
||||||
the performance (SSD).
|
the performance (SSD).
|
||||||
|
|
||||||
As `<device>` it is possible to use more devices, like it's shown in
|
For `<device>`, you can use multiple devices, as is shown in
|
||||||
"Create a new pool with RAID*".
|
"Create a new pool with RAID*".
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
@ -146,8 +146,9 @@ As `<device>` it is possible to use more devices, like it's shown in
|
|||||||
Add cache and log to an existing pool
|
Add cache and log to an existing pool
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
If you have a pool without cache and log. First partition the SSD in
|
You can add cache and log devices to a pool after its creation. In this example,
|
||||||
2 partition with `parted` or `gdisk`
|
we will use a single drive for both cache and log. First, you need to create
|
||||||
|
2 partitions on the SSD with `parted` or `gdisk`
|
||||||
|
|
||||||
.. important:: Always use GPT partition tables.
|
.. important:: Always use GPT partition tables.
|
||||||
|
|
||||||
@ -171,12 +172,12 @@ Changing a failed device
|
|||||||
Changing a failed bootable device
|
Changing a failed bootable device
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
Depending on how Proxmox Backup was installed it is either using `grub` or `systemd-boot`
|
Depending on how Proxmox Backup was installed, it is either using `grub` or
|
||||||
as bootloader.
|
`systemd-boot` as a bootloader.
|
||||||
|
|
||||||
The first steps of copying the partition table, reissuing GUIDs and replacing
|
In either case, the first steps of copying the partition table, reissuing GUIDs
|
||||||
the ZFS partition are the same. To make the system bootable from the new disk,
|
and replacing the ZFS partition are the same. To make the system bootable from
|
||||||
different steps are needed which depend on the bootloader in use.
|
the new disk, different steps are needed which depend on the bootloader in use.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -207,7 +208,7 @@ Usually `grub.cfg` is located in `/boot/grub/grub.cfg`
|
|||||||
# grub-mkconfig -o /path/to/grub.cfg
|
# grub-mkconfig -o /path/to/grub.cfg
|
||||||
|
|
||||||
|
|
||||||
Activate E-Mail Notification
|
Activate e-mail notification
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
ZFS comes with an event daemon, which monitors events generated by the
|
ZFS comes with an event daemon, which monitors events generated by the
|
||||||
@ -219,24 +220,24 @@ and you can install it using `apt-get`:
|
|||||||
|
|
||||||
# apt-get install zfs-zed
|
# apt-get install zfs-zed
|
||||||
|
|
||||||
To activate the daemon it is necessary to edit `/etc/zfs/zed.d/zed.rc` with your
|
To activate the daemon, it is necessary to to uncomment the ZED_EMAIL_ADDR
|
||||||
favorite editor, and uncomment the `ZED_EMAIL_ADDR` setting:
|
setting, in the file `/etc/zfs/zed.d/zed.rc`.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
ZED_EMAIL_ADDR="root"
|
ZED_EMAIL_ADDR="root"
|
||||||
|
|
||||||
Please note Proxmox Backup forwards mails to `root` to the email address
|
Please note that Proxmox Backup forwards mails to `root` to the email address
|
||||||
configured for the root user.
|
configured for the root user.
|
||||||
|
|
||||||
IMPORTANT: The only setting that is required is `ZED_EMAIL_ADDR`. All
|
IMPORTANT: The only setting that is required is `ZED_EMAIL_ADDR`. All
|
||||||
other settings are optional.
|
other settings are optional.
|
||||||
|
|
||||||
Limit ZFS Memory Usage
|
Limit ZFS memory usage
|
||||||
^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
It is good to use at most 50 percent (which is the default) of the
|
It is good to use at most 50 percent (which is the default) of the
|
||||||
system memory for ZFS ARC to prevent performance shortage of the
|
system memory for ZFS ARC, to prevent performance degradation of the
|
||||||
host. Use your preferred editor to change the configuration in
|
host. Use your preferred editor to change the configuration in
|
||||||
`/etc/modprobe.d/zfs.conf` and insert:
|
`/etc/modprobe.d/zfs.conf` and insert:
|
||||||
|
|
||||||
@ -244,27 +245,42 @@ host. Use your preferred editor to change the configuration in
|
|||||||
|
|
||||||
options zfs zfs_arc_max=8589934592
|
options zfs zfs_arc_max=8589934592
|
||||||
|
|
||||||
This example setting limits the usage to 8GB.
|
The above example limits the usage to 8 GiB ('8 * 2^30^').
|
||||||
|
|
||||||
.. IMPORTANT:: If your root file system is ZFS you must update your initramfs every time this value changes:
|
.. IMPORTANT:: In case your desired `zfs_arc_max` value is lower than or equal
|
||||||
|
to `zfs_arc_min` (which defaults to 1/32 of the system memory), `zfs_arc_max`
|
||||||
|
will be ignored. Thus, for it to work in this case, you must set
|
||||||
|
`zfs_arc_min` to at most `zfs_arc_max - 1`. This would require updating the
|
||||||
|
configuration in `/etc/modprobe.d/zfs.conf`, with:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
options zfs zfs_arc_min=8589934591
|
||||||
|
options zfs zfs_arc_max=8589934592
|
||||||
|
|
||||||
|
This example setting limits the usage to 8 GiB ('8 * 2^30^') on
|
||||||
|
systems with more than 256 GiB of total memory, where simply setting
|
||||||
|
`zfs_arc_max` alone would not work.
|
||||||
|
|
||||||
|
.. IMPORTANT:: If your root file system is ZFS, you must update your initramfs
|
||||||
|
every time this value changes.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
# update-initramfs -u
|
# update-initramfs -u
|
||||||
|
|
||||||
|
|
||||||
SWAP on ZFS
|
Swap on ZFS
|
||||||
^^^^^^^^^^^
|
^^^^^^^^^^^
|
||||||
|
|
||||||
Swap-space created on a zvol may generate some troubles, like blocking the
|
Swap-space created on a zvol may cause some issues, such as blocking the
|
||||||
server or generating a high IO load, often seen when starting a Backup
|
server or generating a high IO load, often seen when starting a Backup
|
||||||
to an external Storage.
|
to an external Storage.
|
||||||
|
|
||||||
We strongly recommend to use enough memory, so that you normally do not
|
We strongly recommend using enough memory, so that you normally do not
|
||||||
run into low memory situations. Should you need or want to add swap, it is
|
run into low memory situations. Should you need or want to add swap, it is
|
||||||
preferred to create a partition on a physical disk and use it as swap device.
|
preferred to create a partition on a physical disk and use it as a swap device.
|
||||||
You can leave some space free for this purpose in the advanced options of the
|
You can leave some space free for this purpose in the advanced options of the
|
||||||
installer. Additionally, you can lower the `swappiness` value.
|
installer. Additionally, you can lower the `swappiness` value.
|
||||||
A good value for servers is 10:
|
A good value for servers is 10:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
@ -291,7 +307,7 @@ an editor of your choice and add the following line:
|
|||||||
vm.swappiness = 100 The kernel will swap aggressively.
|
vm.swappiness = 100 The kernel will swap aggressively.
|
||||||
==================== ===============================================================
|
==================== ===============================================================
|
||||||
|
|
||||||
ZFS Compression
|
ZFS compression
|
||||||
^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
To activate compression:
|
To activate compression:
|
||||||
@ -300,10 +316,11 @@ To activate compression:
|
|||||||
# zpool set compression=lz4 <pool>
|
# zpool set compression=lz4 <pool>
|
||||||
|
|
||||||
We recommend using the `lz4` algorithm, since it adds very little CPU overhead.
|
We recommend using the `lz4` algorithm, since it adds very little CPU overhead.
|
||||||
Other algorithms such as `lzjb` and `gzip-N` (where `N` is an integer `1-9` representing
|
Other algorithms such as `lzjb` and `gzip-N` (where `N` is an integer from `1-9`
|
||||||
the compression ratio, 1 is fastest and 9 is best compression) are also available.
|
representing the compression ratio, where 1 is fastest and 9 is best
|
||||||
Depending on the algorithm and how compressible the data is, having compression enabled can even increase
|
compression) are also available. Depending on the algorithm and how
|
||||||
I/O performance.
|
compressible the data is, having compression enabled can even increase I/O
|
||||||
|
performance.
|
||||||
|
|
||||||
You can disable compression at any time with:
|
You can disable compression at any time with:
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
@ -314,26 +331,26 @@ Only new blocks will be affected by this change.
|
|||||||
|
|
||||||
.. _local_zfs_special_device:
|
.. _local_zfs_special_device:
|
||||||
|
|
||||||
ZFS Special Device
|
ZFS special device
|
||||||
^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
Since version 0.8.0 ZFS supports `special` devices. A `special` device in a
|
Since version 0.8.0, ZFS supports `special` devices. A `special` device in a
|
||||||
pool is used to store metadata, deduplication tables, and optionally small
|
pool is used to store metadata, deduplication tables, and optionally small
|
||||||
file blocks.
|
file blocks.
|
||||||
|
|
||||||
A `special` device can improve the speed of a pool consisting of slow spinning
|
A `special` device can improve the speed of a pool consisting of slow spinning
|
||||||
hard disks with a lot of metadata changes. For example workloads that involve
|
hard disks with a lot of metadata changes. For example, workloads that involve
|
||||||
creating, updating or deleting a large number of files will benefit from the
|
creating, updating or deleting a large number of files will benefit from the
|
||||||
presence of a `special` device. ZFS datasets can also be configured to store
|
presence of a `special` device. ZFS datasets can also be configured to store
|
||||||
whole small files on the `special` device which can further improve the
|
small files on the `special` device, which can further improve the
|
||||||
performance. Use fast SSDs for the `special` device.
|
performance. Use fast SSDs for the `special` device.
|
||||||
|
|
||||||
.. IMPORTANT:: The redundancy of the `special` device should match the one of the
|
.. IMPORTANT:: The redundancy of the `special` device should match the one of the
|
||||||
pool, since the `special` device is a point of failure for the whole pool.
|
pool, since the `special` device is a point of failure for the entire pool.
|
||||||
|
|
||||||
.. WARNING:: Adding a `special` device to a pool cannot be undone!
|
.. WARNING:: Adding a `special` device to a pool cannot be undone!
|
||||||
|
|
||||||
Create a pool with `special` device and RAID-1:
|
To create a pool with `special` device and RAID-1:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -346,8 +363,8 @@ Adding a `special` device to an existing pool with RAID-1:
|
|||||||
# zpool add <pool> special mirror <device1> <device2>
|
# zpool add <pool> special mirror <device1> <device2>
|
||||||
|
|
||||||
ZFS datasets expose the `special_small_blocks=<size>` property. `size` can be
|
ZFS datasets expose the `special_small_blocks=<size>` property. `size` can be
|
||||||
`0` to disable storing small file blocks on the `special` device or a power of
|
`0` to disable storing small file blocks on the `special` device, or a power of
|
||||||
two in the range between `512B` to `128K`. After setting the property new file
|
two in the range between `512B` to `128K`. After setting this property, new file
|
||||||
blocks smaller than `size` will be allocated on the `special` device.
|
blocks smaller than `size` will be allocated on the `special` device.
|
||||||
|
|
||||||
.. IMPORTANT:: If the value for `special_small_blocks` is greater than or equal to
|
.. IMPORTANT:: If the value for `special_small_blocks` is greater than or equal to
|
||||||
@ -355,10 +372,10 @@ blocks smaller than `size` will be allocated on the `special` device.
|
|||||||
the `special` device, so be careful!
|
the `special` device, so be careful!
|
||||||
|
|
||||||
Setting the `special_small_blocks` property on a pool will change the default
|
Setting the `special_small_blocks` property on a pool will change the default
|
||||||
value of that property for all child ZFS datasets (for example all containers
|
value of that property for all child ZFS datasets (for example, all containers
|
||||||
in the pool will opt in for small file blocks).
|
in the pool will opt in for small file blocks).
|
||||||
|
|
||||||
Opt in for all file smaller than 4K-blocks pool-wide:
|
Opt in for all files smaller than 4K-blocks pool-wide:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -379,10 +396,15 @@ Opt out from small file blocks for a single dataset:
|
|||||||
Troubleshooting
|
Troubleshooting
|
||||||
^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
Corrupted cachefile
|
Corrupt cache file
|
||||||
|
""""""""""""""""""
|
||||||
|
|
||||||
In case of a corrupted ZFS cachefile, some volumes may not be mounted during
|
`zfs-import-cache.service` imports ZFS pools using the ZFS cache file. If this
|
||||||
boot until mounted manually later.
|
file becomes corrupted, the service won't be able to import the pools that it's
|
||||||
|
unable to read from it.
|
||||||
|
|
||||||
|
As a result, in case of a corrupted ZFS cache file, some volumes may not be
|
||||||
|
mounted during boot and must be mounted manually later.
|
||||||
|
|
||||||
For each pool, run:
|
For each pool, run:
|
||||||
|
|
||||||
@ -390,16 +412,13 @@ For each pool, run:
|
|||||||
|
|
||||||
# zpool set cachefile=/etc/zfs/zpool.cache POOLNAME
|
# zpool set cachefile=/etc/zfs/zpool.cache POOLNAME
|
||||||
|
|
||||||
and afterwards update the `initramfs` by running:
|
then, update the `initramfs` by running:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
# update-initramfs -u -k all
|
# update-initramfs -u -k all
|
||||||
|
|
||||||
and finally reboot your node.
|
and finally, reboot the node.
|
||||||
|
|
||||||
Sometimes the ZFS cachefile can get corrupted, and `zfs-import-cache.service`
|
|
||||||
doesn't import the pools that aren't present in the cachefile.
|
|
||||||
|
|
||||||
Another workaround to this problem is enabling the `zfs-import-scan.service`,
|
Another workaround to this problem is enabling the `zfs-import-scan.service`,
|
||||||
which searches and imports pools via device scanning (usually slower).
|
which searches and imports pools via device scanning (usually slower).
|
||||||
|
@ -34,17 +34,7 @@
|
|||||||
</style>
|
</style>
|
||||||
<link rel="stylesheet" type="text/css" href="font-awesome/css/font-awesome.css"/>
|
<link rel="stylesheet" type="text/css" href="font-awesome/css/font-awesome.css"/>
|
||||||
<script type="text/javascript" src="extjs/ext-all.js"></script>
|
<script type="text/javascript" src="extjs/ext-all.js"></script>
|
||||||
|
<script type="text/javascript" src="lto-barcode-generator.js"></script>
|
||||||
<script type="text/javascript" src="code39.js"></script>
|
|
||||||
<script type="text/javascript" src="prefix-field.js"></script>
|
|
||||||
<script type="text/javascript" src="label-style.js"></script>
|
|
||||||
<script type="text/javascript" src="tape-type.js"></script>
|
|
||||||
<script type="text/javascript" src="paper-size.js"></script>
|
|
||||||
<script type="text/javascript" src="page-layout.js"></script>
|
|
||||||
<script type="text/javascript" src="page-calibration.js"></script>
|
|
||||||
<script type="text/javascript" src="label-list.js"></script>
|
|
||||||
<script type="text/javascript" src="label-setup.js"></script>
|
|
||||||
<script type="text/javascript" src="lto-barcode.js"></script>
|
|
||||||
</head>
|
</head>
|
||||||
<body>
|
<body>
|
||||||
</body>
|
</body>
|
||||||
|
@ -1,7 +1,5 @@
|
|||||||
// FIXME: HACK! Makes scrolling in number spinner work again. fixed in ExtJS >= 6.1
|
// for toolkit.js
|
||||||
if (Ext.isFirefox) {
|
function gettext(val) { return val; };
|
||||||
Ext.$eventNameMap.DOMMouseScroll = 'DOMMouseScroll';
|
|
||||||
}
|
|
||||||
|
|
||||||
function draw_labels(target_id, label_list, page_layout, calibration) {
|
function draw_labels(target_id, label_list, page_layout, calibration) {
|
||||||
let max_labels = compute_max_labels(page_layout);
|
let max_labels = compute_max_labels(page_layout);
|
||||||
|
@ -14,15 +14,15 @@ following retention options are available:
|
|||||||
|
|
||||||
``keep-hourly <N>``
|
``keep-hourly <N>``
|
||||||
Keep backups for the last ``<N>`` hours. If there is more than one
|
Keep backups for the last ``<N>`` hours. If there is more than one
|
||||||
backup for a single hour, only the latest is kept.
|
backup for a single hour, only the latest is retained.
|
||||||
|
|
||||||
``keep-daily <N>``
|
``keep-daily <N>``
|
||||||
Keep backups for the last ``<N>`` days. If there is more than one
|
Keep backups for the last ``<N>`` days. If there is more than one
|
||||||
backup for a single day, only the latest is kept.
|
backup for a single day, only the latest is retained.
|
||||||
|
|
||||||
``keep-weekly <N>``
|
``keep-weekly <N>``
|
||||||
Keep backups for the last ``<N>`` weeks. If there is more than one
|
Keep backups for the last ``<N>`` weeks. If there is more than one
|
||||||
backup for a single week, only the latest is kept.
|
backup for a single week, only the latest is retained.
|
||||||
|
|
||||||
.. note:: Weeks start on Monday and end on Sunday. The software
|
.. note:: Weeks start on Monday and end on Sunday. The software
|
||||||
uses the `ISO week date`_ system and handles weeks at
|
uses the `ISO week date`_ system and handles weeks at
|
||||||
@ -30,17 +30,17 @@ following retention options are available:
|
|||||||
|
|
||||||
``keep-monthly <N>``
|
``keep-monthly <N>``
|
||||||
Keep backups for the last ``<N>`` months. If there is more than one
|
Keep backups for the last ``<N>`` months. If there is more than one
|
||||||
backup for a single month, only the latest is kept.
|
backup for a single month, only the latest is retained.
|
||||||
|
|
||||||
``keep-yearly <N>``
|
``keep-yearly <N>``
|
||||||
Keep backups for the last ``<N>`` years. If there is more than one
|
Keep backups for the last ``<N>`` years. If there is more than one
|
||||||
backup for a single year, only the latest is kept.
|
backup for a single year, only the latest is retained.
|
||||||
|
|
||||||
The retention options are processed in the order given above. Each option
|
The retention options are processed in the order given above. Each option
|
||||||
only covers backups within its time period. The next option does not take care
|
only covers backups within its time period. The next option does not take care
|
||||||
of already covered backups. It will only consider older backups.
|
of already covered backups. It will only consider older backups.
|
||||||
|
|
||||||
Unfinished and incomplete backups will be removed by the prune command unless
|
Unfinished and incomplete backups will be removed by the prune command, unless
|
||||||
they are newer than the last successful backup. In this case, the last failed
|
they are newer than the last successful backup. In this case, the last failed
|
||||||
backup is retained.
|
backup is retained.
|
||||||
|
|
||||||
@ -48,7 +48,7 @@ Prune Simulator
|
|||||||
^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
You can use the built-in `prune simulator <prune-simulator/index.html>`_
|
You can use the built-in `prune simulator <prune-simulator/index.html>`_
|
||||||
to explore the effect of different retetion options with various backup
|
to explore the effect of different retention options with various backup
|
||||||
schedules.
|
schedules.
|
||||||
|
|
||||||
Manual Pruning
|
Manual Pruning
|
||||||
@ -59,10 +59,10 @@ Manual Pruning
|
|||||||
:align: right
|
:align: right
|
||||||
:alt: Prune and garbage collection options
|
:alt: Prune and garbage collection options
|
||||||
|
|
||||||
To access pruning functionality for a specific backup group, you can use the
|
To manually prune a specific backup group, you can use
|
||||||
prune command line option discussed in :ref:`backup-pruning`, or navigate to
|
``proxmox-backup-client``'s ``prune`` subcommand, discussed in
|
||||||
the **Content** tab of the datastore and click the scissors icon in the
|
:ref:`backup-pruning`, or navigate to the **Content** tab of the datastore and
|
||||||
**Actions** column of the relevant backup group.
|
click the scissors icon in the **Actions** column of the relevant backup group.
|
||||||
|
|
||||||
Prune Schedules
|
Prune Schedules
|
||||||
^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^
|
||||||
@ -81,7 +81,7 @@ Retention Settings Example
|
|||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
The backup frequency and retention of old backups may depend on how often data
|
The backup frequency and retention of old backups may depend on how often data
|
||||||
changes, and how important an older state may be, in a specific work load.
|
changes and how important an older state may be in a specific workload.
|
||||||
When backups act as a company's document archive, there may also be legal
|
When backups act as a company's document archive, there may also be legal
|
||||||
requirements for how long backup snapshots must be kept.
|
requirements for how long backup snapshots must be kept.
|
||||||
|
|
||||||
@ -125,8 +125,8 @@ start garbage collection on an entire datastore and the ``status`` subcommand to
|
|||||||
see attributes relating to the :ref:`garbage collection <client_garbage-collection>`.
|
see attributes relating to the :ref:`garbage collection <client_garbage-collection>`.
|
||||||
|
|
||||||
This functionality can also be accessed in the GUI, by navigating to **Prune &
|
This functionality can also be accessed in the GUI, by navigating to **Prune &
|
||||||
GC** from the top panel. From here, you can edit the schedule at which garbage
|
GC** from the top panel of a datastore. From here, you can edit the schedule at
|
||||||
collection runs and manually start the operation.
|
which garbage collection runs and manually start the operation.
|
||||||
|
|
||||||
|
|
||||||
.. _maintenance_verification:
|
.. _maintenance_verification:
|
||||||
@ -139,13 +139,13 @@ Verification
|
|||||||
:align: right
|
:align: right
|
||||||
:alt: Adding a verify job
|
:alt: Adding a verify job
|
||||||
|
|
||||||
Proxmox Backup offers various verification options to ensure that backup data is
|
Proxmox Backup Server offers various verification options to ensure that backup
|
||||||
intact. Verification is generally carried out through the creation of verify
|
data is intact. Verification is generally carried out through the creation of
|
||||||
jobs. These are scheduled tasks that run verification at a given interval (see
|
verify jobs. These are scheduled tasks that run verification at a given interval
|
||||||
:ref:`calendar-event-scheduling`). With these, you can set whether already verified
|
(see :ref:`calendar-event-scheduling`). With these, you can also set whether
|
||||||
snapshots are ignored, as well as set a time period, after which verified jobs
|
already verified snapshots are ignored, as well as set a time period, after
|
||||||
are checked again. The interface for creating verify jobs can be found under the
|
which snapshots are checked again. The interface for creating verify jobs can be
|
||||||
**Verify Jobs** tab of the datastore.
|
found under the **Verify Jobs** tab of the datastore.
|
||||||
|
|
||||||
.. Note:: It is recommended that you reverify all backups at least monthly, even
|
.. Note:: It is recommended that you reverify all backups at least monthly, even
|
||||||
if a previous verification was successful. This is because physical drives
|
if a previous verification was successful. This is because physical drives
|
||||||
@ -158,9 +158,9 @@ are checked again. The interface for creating verify jobs can be found under the
|
|||||||
data.
|
data.
|
||||||
|
|
||||||
Aside from using verify jobs, you can also run verification manually on entire
|
Aside from using verify jobs, you can also run verification manually on entire
|
||||||
datastores, backup groups, or snapshots. To do this, navigate to the **Content**
|
datastores, backup groups or snapshots. To do this, navigate to the **Content**
|
||||||
tab of the datastore and either click *Verify All*, or select the *V.* icon from
|
tab of the datastore and either click *Verify All* or select the *V.* icon from
|
||||||
the *Actions* column in the table.
|
the **Actions** column in the table.
|
||||||
|
|
||||||
.. _maintenance_notification:
|
.. _maintenance_notification:
|
||||||
|
|
||||||
@ -170,8 +170,8 @@ Notifications
|
|||||||
Proxmox Backup Server can send you notification emails about automatically
|
Proxmox Backup Server can send you notification emails about automatically
|
||||||
scheduled verification, garbage-collection and synchronization tasks results.
|
scheduled verification, garbage-collection and synchronization tasks results.
|
||||||
|
|
||||||
By default, notifications are send to the email address configured for the
|
By default, notifications are sent to the email address configured for the
|
||||||
`root@pam` user. You can set that user for each datastore.
|
`root@pam` user. You can instead set this user for each datastore.
|
||||||
|
|
||||||
You can also change the level of notification received per task type, the
|
You can also change the level of notification received per task type, the
|
||||||
following options are available:
|
following options are available:
|
||||||
@ -179,6 +179,6 @@ following options are available:
|
|||||||
* Always: send a notification for any scheduled task, independent of the
|
* Always: send a notification for any scheduled task, independent of the
|
||||||
outcome
|
outcome
|
||||||
|
|
||||||
* Errors: send a notification for any scheduled task resulting in an error
|
* Errors: send a notification for any scheduled task that results in an error
|
||||||
|
|
||||||
* Never: do not send any notification at all
|
* Never: do not send any notification at all
|
||||||
|
@ -17,8 +17,8 @@ configuration information for remotes is stored in the file
|
|||||||
:align: right
|
:align: right
|
||||||
:alt: Add a remote
|
:alt: Add a remote
|
||||||
|
|
||||||
To add a remote, you need its hostname or IP, a userid and password on the
|
To add a remote, you need its hostname or IP address, a userid and password on
|
||||||
remote, and its certificate fingerprint. To get the fingerprint, use the
|
the remote, and its certificate fingerprint. To get the fingerprint, use the
|
||||||
``proxmox-backup-manager cert info`` command on the remote, or navigate to
|
``proxmox-backup-manager cert info`` command on the remote, or navigate to
|
||||||
**Dashboard** in the remote's web interface and select **Show Fingerprint**.
|
**Dashboard** in the remote's web interface and select **Show Fingerprint**.
|
||||||
|
|
||||||
@ -60,12 +60,13 @@ Sync Jobs
|
|||||||
|
|
||||||
Sync jobs are configured to pull the contents of a datastore on a **Remote** to
|
Sync jobs are configured to pull the contents of a datastore on a **Remote** to
|
||||||
a local datastore. You can manage sync jobs in the web interface, from the
|
a local datastore. You can manage sync jobs in the web interface, from the
|
||||||
**Sync Jobs** tab of the datastore which you'd like to set one up for, or using
|
**Sync Jobs** tab of the **Datastore** panel or from that of the Datastore
|
||||||
the ``proxmox-backup-manager sync-job`` command. The configuration information
|
itself. Alternatively, you can manage them with the ``proxmox-backup-manager
|
||||||
for sync jobs is stored at ``/etc/proxmox-backup/sync.cfg``. To create a new
|
sync-job`` command. The configuration information for sync jobs is stored at
|
||||||
sync job, click the add button in the GUI, or use the ``create`` subcommand.
|
``/etc/proxmox-backup/sync.cfg``. To create a new sync job, click the add button
|
||||||
After creating a sync job, you can either start it manually from the GUI or
|
in the GUI, or use the ``create`` subcommand. After creating a sync job, you can
|
||||||
provide it with a schedule (see :ref:`calendar-event-scheduling`) to run regularly.
|
either start it manually from the GUI or provide it with a schedule (see
|
||||||
|
:ref:`calendar-event-scheduling`) to run regularly.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -79,17 +80,48 @@ provide it with a schedule (see :ref:`calendar-event-scheduling`) to run regular
|
|||||||
└────────────┴───────┴────────┴──────────────┴───────────┴─────────┘
|
└────────────┴───────┴────────┴──────────────┴───────────┴─────────┘
|
||||||
# proxmox-backup-manager sync-job remove pbs2-local
|
# proxmox-backup-manager sync-job remove pbs2-local
|
||||||
|
|
||||||
For setting up sync jobs, the configuring user needs the following permissions:
|
To set up sync jobs, the configuring user needs the following permissions:
|
||||||
|
|
||||||
#. ``Remote.Read`` on the ``/remote/{remote}/{remote-store}`` path
|
#. ``Remote.Read`` on the ``/remote/{remote}/{remote-store}`` path
|
||||||
#. at least ``Datastore.Backup`` on the local target datastore (``/datastore/{store}``)
|
#. At least ``Datastore.Backup`` on the local target datastore (``/datastore/{store}``)
|
||||||
|
|
||||||
If the ``remove-vanished`` option is set, ``Datastore.Prune`` is required on
|
|
||||||
the local datastore as well. If the ``owner`` option is not set (defaulting to
|
|
||||||
``root@pam``) or set to something other than the configuring user,
|
|
||||||
``Datastore.Modify`` is required as well.
|
|
||||||
|
|
||||||
.. note:: A sync job can only sync backup groups that the configured remote's
|
.. note:: A sync job can only sync backup groups that the configured remote's
|
||||||
user/API token can read. If a remote is configured with a user/API token that
|
user/API token can read. If a remote is configured with a user/API token that
|
||||||
only has ``Datastore.Backup`` privileges, only the limited set of accessible
|
only has ``Datastore.Backup`` privileges, only the limited set of accessible
|
||||||
snapshots owned by that user/API token can be synced.
|
snapshots owned by that user/API token can be synced.
|
||||||
|
|
||||||
|
If the ``remove-vanished`` option is set, ``Datastore.Prune`` is required on
|
||||||
|
the local datastore as well. If the ``owner`` option is not set (defaulting to
|
||||||
|
``root@pam``) or is set to something other than the configuring user,
|
||||||
|
``Datastore.Modify`` is required as well.
|
||||||
|
|
||||||
|
If the ``group-filter`` option is set, only backup groups matching at least one
|
||||||
|
of the specified criteria are synced. The available criteria are:
|
||||||
|
|
||||||
|
* backup type, for example to only sync groups of the `ct` (Container) type:
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-manager sync-job update ID --group-filter type:ct
|
||||||
|
* full group identifier
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-manager sync-job update ID --group-filter group:vm/100
|
||||||
|
* regular expression matched against the full group identifier
|
||||||
|
.. todo:: add example for regex
|
||||||
|
|
||||||
|
The same filter is applied to local groups for handling of the
|
||||||
|
``remove-vanished`` option.
|
||||||
|
|
||||||
|
.. note:: The ``protected`` flag of remote backup snapshots will not be synced.
|
||||||
|
|
||||||
|
Bandwidth Limit
|
||||||
|
^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
|
Syncing datastores to an archive can produce lots of traffic and impact other
|
||||||
|
users of the network. So, to avoid network or storage congetsion you can limit
|
||||||
|
the bandwith of the sync job by setting the ``rate-in`` option either in the
|
||||||
|
web interface or using the ``proxmox-backup-manager`` command-line tool:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-manager sync-job update ID --rate-in 20MiB
|
||||||
|
@ -82,9 +82,12 @@ is:
|
|||||||
.. note:: This command and corresponding GUI button rely on the ``ifreload``
|
.. note:: This command and corresponding GUI button rely on the ``ifreload``
|
||||||
command, from the package ``ifupdown2``. This package is included within the
|
command, from the package ``ifupdown2``. This package is included within the
|
||||||
Proxmox Backup Server installation, however, you may have to install it yourself,
|
Proxmox Backup Server installation, however, you may have to install it yourself,
|
||||||
if you have installed Proxmox Backup Server on top of Debian or Proxmox VE.
|
if you have installed Proxmox Backup Server on top of Debian or a Proxmox VE
|
||||||
|
version prior to version 7.
|
||||||
|
|
||||||
You can also configure DNS settings, from the **DNS** section
|
You can also configure DNS settings, from the **DNS** section
|
||||||
of **Configuration** or by using the ``dns`` subcommand of
|
of **Configuration** or by using the ``dns`` subcommand of
|
||||||
``proxmox-backup-manager``.
|
``proxmox-backup-manager``.
|
||||||
|
|
||||||
|
|
||||||
|
.. include:: traffic-control.rst
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
Most commands producing output supports the ``--output-format``
|
Most commands that produce output support the ``--output-format``
|
||||||
parameter. It accepts the following values:
|
parameter. This accepts the following values:
|
||||||
|
|
||||||
:``text``: Text format (default). Structured data is rendered as a table.
|
:``text``: Text format (default). Structured data is rendered as a table.
|
||||||
|
|
||||||
|
@ -17,15 +17,13 @@ update``.
|
|||||||
.. code-block:: sources.list
|
.. code-block:: sources.list
|
||||||
:caption: File: ``/etc/apt/sources.list``
|
:caption: File: ``/etc/apt/sources.list``
|
||||||
|
|
||||||
deb http://ftp.debian.org/debian buster main contrib
|
deb http://ftp.debian.org/debian bullseye main contrib
|
||||||
deb http://ftp.debian.org/debian buster-updates main contrib
|
deb http://ftp.debian.org/debian bullseye-updates main contrib
|
||||||
|
|
||||||
# security updates
|
# security updates
|
||||||
deb http://security.debian.org/debian-security buster/updates main contrib
|
deb http://security.debian.org/debian-security bullseye-security main contrib
|
||||||
|
|
||||||
|
|
||||||
.. FIXME for 7.0: change security update suite to bullseye-security
|
|
||||||
|
|
||||||
In addition, you need a package repository from Proxmox to get Proxmox Backup
|
In addition, you need a package repository from Proxmox to get Proxmox Backup
|
||||||
updates.
|
updates.
|
||||||
|
|
||||||
@ -45,31 +43,21 @@ key with the following commands:
|
|||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
# wget http://download.proxmox.com/debian/proxmox-ve-release-6.x.gpg -O /etc/apt/trusted.gpg.d/proxmox-ve-release-6.x.gpg
|
# wget https://enterprise.proxmox.com/debian/proxmox-release-bullseye.gpg -O /etc/apt/trusted.gpg.d/proxmox-release-bullseye.gpg
|
||||||
|
|
||||||
Verify the SHA512 checksum afterwards with:
|
Verify the SHA512 checksum afterwards with the expected output below:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
# sha512sum /etc/apt/trusted.gpg.d/proxmox-ve-release-6.x.gpg
|
# sha512sum /etc/apt/trusted.gpg.d/proxmox-release-bullseye.gpg
|
||||||
|
7fb03ec8a1675723d2853b84aa4fdb49a46a3bb72b9951361488bfd19b29aab0a789a4f8c7406e71a69aabbc727c936d3549731c4659ffa1a08f44db8fdcebfa /etc/apt/trusted.gpg.d/proxmox-release-bullseye.gpg
|
||||||
|
|
||||||
The output should be:
|
and the md5sum, with the expected output below:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
acca6f416917e8e11490a08a1e2842d500b3a5d9f322c6319db0927b2901c3eae23cfb5cd5df6facf2b57399d3cfa52ad7769ebdd75d9b204549ca147da52626 /etc/apt/trusted.gpg.d/proxmox-ve-release-6.x.gpg
|
# md5sum /etc/apt/trusted.gpg.d/proxmox-release-bullseye.gpg
|
||||||
|
bcc35c7173e0845c0d6ad6470b70f50e /etc/apt/trusted.gpg.d/proxmox-release-bullseye.gpg
|
||||||
and the md5sum:
|
|
||||||
|
|
||||||
.. code-block:: console
|
|
||||||
|
|
||||||
# md5sum /etc/apt/trusted.gpg.d/proxmox-ve-release-6.x.gpg
|
|
||||||
|
|
||||||
Here, the output should be:
|
|
||||||
|
|
||||||
.. code-block:: console
|
|
||||||
|
|
||||||
f3f6c5a3a67baf38ad178e5ff1ee270c /etc/apt/trusted.gpg.d/proxmox-ve-release-6.x.gpg
|
|
||||||
|
|
||||||
.. _sysadmin_package_repos_enterprise:
|
.. _sysadmin_package_repos_enterprise:
|
||||||
|
|
||||||
@ -84,7 +72,7 @@ enabled by default:
|
|||||||
.. code-block:: sources.list
|
.. code-block:: sources.list
|
||||||
:caption: File: ``/etc/apt/sources.list.d/pbs-enterprise.list``
|
:caption: File: ``/etc/apt/sources.list.d/pbs-enterprise.list``
|
||||||
|
|
||||||
deb https://enterprise.proxmox.com/debian/pbs buster pbs-enterprise
|
deb https://enterprise.proxmox.com/debian/pbs bullseye pbs-enterprise
|
||||||
|
|
||||||
|
|
||||||
To never miss important security fixes, the superuser (``root@pam`` user) is
|
To never miss important security fixes, the superuser (``root@pam`` user) is
|
||||||
@ -114,15 +102,15 @@ We recommend to configure this repository in ``/etc/apt/sources.list``.
|
|||||||
.. code-block:: sources.list
|
.. code-block:: sources.list
|
||||||
:caption: File: ``/etc/apt/sources.list``
|
:caption: File: ``/etc/apt/sources.list``
|
||||||
|
|
||||||
deb http://ftp.debian.org/debian buster main contrib
|
deb http://ftp.debian.org/debian bullseye main contrib
|
||||||
deb http://ftp.debian.org/debian buster-updates main contrib
|
deb http://ftp.debian.org/debian bullseye-updates main contrib
|
||||||
|
|
||||||
# PBS pbs-no-subscription repository provided by proxmox.com,
|
# PBS pbs-no-subscription repository provided by proxmox.com,
|
||||||
# NOT recommended for production use
|
# NOT recommended for production use
|
||||||
deb http://download.proxmox.com/debian/pbs buster pbs-no-subscription
|
deb http://download.proxmox.com/debian/pbs bullseye pbs-no-subscription
|
||||||
|
|
||||||
# security updates
|
# security updates
|
||||||
deb http://security.debian.org/debian-security buster/updates main contrib
|
deb http://security.debian.org/debian-security bullseye-security main contrib
|
||||||
|
|
||||||
|
|
||||||
`Proxmox Backup`_ Test Repository
|
`Proxmox Backup`_ Test Repository
|
||||||
@ -140,7 +128,7 @@ You can access this repository by adding the following line to
|
|||||||
.. code-block:: sources.list
|
.. code-block:: sources.list
|
||||||
:caption: sources.list entry for ``pbstest``
|
:caption: sources.list entry for ``pbstest``
|
||||||
|
|
||||||
deb http://download.proxmox.com/debian/pbs buster pbstest
|
deb http://download.proxmox.com/debian/pbs bullseye pbstest
|
||||||
|
|
||||||
.. _package_repositories_client_only:
|
.. _package_repositories_client_only:
|
||||||
|
|
||||||
@ -161,6 +149,26 @@ APT-based Proxmox Backup Client Repository
|
|||||||
For modern Linux distributions using `apt` as package manager, like all Debian
|
For modern Linux distributions using `apt` as package manager, like all Debian
|
||||||
and Ubuntu Derivative do, you may be able to use the APT-based repository.
|
and Ubuntu Derivative do, you may be able to use the APT-based repository.
|
||||||
|
|
||||||
|
In order to configure this repository you need to first :ref:`setup the Proxmox
|
||||||
|
release key <package_repos_secure_apt>`. After that, add the repository URL to
|
||||||
|
the APT sources lists.
|
||||||
|
|
||||||
|
**Repositories for Debian 11 (Bullseye) based releases**
|
||||||
|
|
||||||
|
This repository is tested with:
|
||||||
|
|
||||||
|
- Debian Bullseye
|
||||||
|
|
||||||
|
Edit the file ``/etc/apt/sources.list.d/pbs-client.list`` and add the following
|
||||||
|
snipped
|
||||||
|
|
||||||
|
.. code-block:: sources.list
|
||||||
|
:caption: File: ``/etc/apt/sources.list``
|
||||||
|
|
||||||
|
deb http://download.proxmox.com/debian/pbs-client bullseye main
|
||||||
|
|
||||||
|
**Repositories for Debian 10 (Buster) based releases**
|
||||||
|
|
||||||
This repository is tested with:
|
This repository is tested with:
|
||||||
|
|
||||||
- Debian Buster
|
- Debian Buster
|
||||||
@ -168,9 +176,6 @@ This repository is tested with:
|
|||||||
|
|
||||||
It may work with older, and should work with more recent released versions.
|
It may work with older, and should work with more recent released versions.
|
||||||
|
|
||||||
In order to configure this repository you need to first :ref:`setup the Proxmox
|
|
||||||
release key <package_repos_secure_apt>`. After that, add the repository URL to
|
|
||||||
the APT sources lists.
|
|
||||||
Edit the file ``/etc/apt/sources.list.d/pbs-client.list`` and add the following
|
Edit the file ``/etc/apt/sources.list.d/pbs-client.list`` and add the following
|
||||||
snipped
|
snipped
|
||||||
|
|
||||||
@ -178,3 +183,19 @@ snipped
|
|||||||
:caption: File: ``/etc/apt/sources.list``
|
:caption: File: ``/etc/apt/sources.list``
|
||||||
|
|
||||||
deb http://download.proxmox.com/debian/pbs-client buster main
|
deb http://download.proxmox.com/debian/pbs-client buster main
|
||||||
|
|
||||||
|
.. _node_options_http_proxy:
|
||||||
|
|
||||||
|
Repository Access Behind HTTP Proxy
|
||||||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
|
Some setups have restricted access to the internet, sometimes only through a
|
||||||
|
central proxy. You can setup a HTTP proxy through the Proxmox Backup Server's
|
||||||
|
web-interface in the `Configuration -> Authentication` tab.
|
||||||
|
|
||||||
|
Once configured this proxy will be used for apt network requests and for
|
||||||
|
checking a Proxmox Backup Server support subscription.
|
||||||
|
|
||||||
|
Standard HTTP proxy configurations are accepted, `[http://]<host>[:port]` where
|
||||||
|
the `<host>` part may include an authorization, for example:
|
||||||
|
`http://user:pass@proxy.example.org:12345`
|
||||||
|
14
docs/proxmox-backup-debug/description.rst
Normal file
14
docs/proxmox-backup-debug/description.rst
Normal file
@ -0,0 +1,14 @@
|
|||||||
|
Implements debugging functionality to inspect Proxmox Backup datastore
|
||||||
|
files, verify the integrity of chunks.
|
||||||
|
|
||||||
|
Also contains an 'api' subcommand where arbitrary api paths can be called
|
||||||
|
(get/create/set/delete) as well as display their parameters (usage) and
|
||||||
|
their child-links (ls).
|
||||||
|
|
||||||
|
By default, it connects to the proxmox-backup-proxy on localhost via https,
|
||||||
|
but by setting the environment variable `PROXMOX_DEBUG_API_CODE` to `1` the
|
||||||
|
tool directly calls the corresponding code.
|
||||||
|
|
||||||
|
.. WARNING:: Using `PROXMOX_DEBUG_API_CODE` can be dangerous and is only intended
|
||||||
|
for debugging purposes. It is not intended for use on a production system.
|
||||||
|
|
33
docs/proxmox-backup-debug/man1.rst
Normal file
33
docs/proxmox-backup-debug/man1.rst
Normal file
@ -0,0 +1,33 @@
|
|||||||
|
==========================
|
||||||
|
proxmox-backup-debug
|
||||||
|
==========================
|
||||||
|
|
||||||
|
.. include:: ../epilog.rst
|
||||||
|
|
||||||
|
-------------------------------------------------------------
|
||||||
|
Debugging command line tool for Backup and Restore
|
||||||
|
-------------------------------------------------------------
|
||||||
|
|
||||||
|
:Author: |AUTHOR|
|
||||||
|
:Version: Version |VERSION|
|
||||||
|
:Manual section: 1
|
||||||
|
|
||||||
|
|
||||||
|
Synopsis
|
||||||
|
==========
|
||||||
|
|
||||||
|
.. include:: synopsis.rst
|
||||||
|
|
||||||
|
Common Options
|
||||||
|
==============
|
||||||
|
|
||||||
|
.. include:: ../output-format.rst
|
||||||
|
|
||||||
|
|
||||||
|
Description
|
||||||
|
============
|
||||||
|
|
||||||
|
.. include:: description.rst
|
||||||
|
|
||||||
|
|
||||||
|
.. include:: ../pbs-copyright.rst
|
@ -1,5 +1,5 @@
|
|||||||
This daemon exposes the whole Proxmox Backup Server API on TCP port
|
This daemon exposes the whole Proxmox Backup Server API on TCP port
|
||||||
8007 using HTTPS. It runs as user ``backup`` and has very limited
|
8007 using HTTPS. It runs as user ``backup`` and has very limited
|
||||||
permissions. Operation requiring more permissions are forwarded to
|
permissions. Operations requiring more permissions are forwarded to
|
||||||
the local ``proxmox-backup`` service.
|
the local ``proxmox-backup`` service.
|
||||||
|
|
||||||
|
@ -1,7 +1,5 @@
|
|||||||
// FIXME: HACK! Makes scrolling in number spinner work again. fixed in ExtJS >= 6.1
|
// for Toolkit.js
|
||||||
if (Ext.isFirefox) {
|
function gettext(val) { return val; };
|
||||||
Ext.$eventNameMap.DOMMouseScroll = 'DOMMouseScroll';
|
|
||||||
}
|
|
||||||
|
|
||||||
Ext.onReady(function() {
|
Ext.onReady(function() {
|
||||||
const NOW = new Date();
|
const NOW = new Date();
|
||||||
@ -37,7 +35,6 @@ Ext.onReady(function() {
|
|||||||
|
|
||||||
editable: true,
|
editable: true,
|
||||||
|
|
||||||
displayField: 'text',
|
|
||||||
valueField: 'value',
|
valueField: 'value',
|
||||||
queryMode: 'local',
|
queryMode: 'local',
|
||||||
|
|
@ -3,8 +3,8 @@
|
|||||||
`Proxmox VE`_ Integration
|
`Proxmox VE`_ Integration
|
||||||
-------------------------
|
-------------------------
|
||||||
|
|
||||||
A Proxmox Backup Server can be integrated into a Proxmox VE setup by adding the
|
Proxmox Backup Server can be integrated into a Proxmox VE standalone or cluster
|
||||||
former as a storage in a Proxmox VE standalone or cluster setup.
|
setup, by adding it as a storage in Proxmox VE.
|
||||||
|
|
||||||
See also the `Proxmox VE Storage - Proxmox Backup Server
|
See also the `Proxmox VE Storage - Proxmox Backup Server
|
||||||
<https://pve.proxmox.com/pve-docs/pve-admin-guide.html#storage_pbs>`_ section
|
<https://pve.proxmox.com/pve-docs/pve-admin-guide.html#storage_pbs>`_ section
|
||||||
@ -14,8 +14,8 @@ of the Proxmox VE Administration Guide for Proxmox VE specific documentation.
|
|||||||
Using the Proxmox VE Web-Interface
|
Using the Proxmox VE Web-Interface
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
Proxmox VE has native API and web-interface integration of Proxmox Backup
|
Proxmox VE has native API and web interface integration of Proxmox Backup
|
||||||
Server since the `Proxmox VE 6.3 release
|
Server as of `Proxmox VE 6.3
|
||||||
<https://pve.proxmox.com/wiki/Roadmap#Proxmox_VE_6.3>`_.
|
<https://pve.proxmox.com/wiki/Roadmap#Proxmox_VE_6.3>`_.
|
||||||
|
|
||||||
A Proxmox Backup Server can be added under ``Datacenter -> Storage``.
|
A Proxmox Backup Server can be added under ``Datacenter -> Storage``.
|
||||||
@ -24,8 +24,8 @@ Using the Proxmox VE Command-Line
|
|||||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
You need to define a new storage with type 'pbs' on your `Proxmox VE`_
|
You need to define a new storage with type 'pbs' on your `Proxmox VE`_
|
||||||
node. The following example uses ``store2`` as storage name, and
|
node. The following example uses ``store2`` as the storage's name, and
|
||||||
assumes the server address is ``localhost``, and you want to connect
|
assumes the server address is ``localhost`` and you want to connect
|
||||||
as ``user1@pbs``.
|
as ``user1@pbs``.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
@ -33,7 +33,7 @@ as ``user1@pbs``.
|
|||||||
# pvesm add pbs store2 --server localhost --datastore store2
|
# pvesm add pbs store2 --server localhost --datastore store2
|
||||||
# pvesm set store2 --username user1@pbs --password <secret>
|
# pvesm set store2 --username user1@pbs --password <secret>
|
||||||
|
|
||||||
.. note:: If you would rather not pass your password as plain text, you can pass
|
.. note:: If you would rather not enter your password as plain text, you can pass
|
||||||
the ``--password`` parameter, without any arguments. This will cause the
|
the ``--password`` parameter, without any arguments. This will cause the
|
||||||
program to prompt you for a password upon entering the command.
|
program to prompt you for a password upon entering the command.
|
||||||
|
|
||||||
@ -53,7 +53,7 @@ relationship:
|
|||||||
|
|
||||||
# pvesm set store2 --fingerprint 64:d3:ff:3a:50:38:53:5a:9b:f7:50:...:ab:fe
|
# pvesm set store2 --fingerprint 64:d3:ff:3a:50:38:53:5a:9b:f7:50:...:ab:fe
|
||||||
|
|
||||||
After that you should be able to see storage status with:
|
After that, you should be able to view storage status with:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
|
@ -1,12 +1,12 @@
|
|||||||
``pxar`` is a command line utility to create and manipulate archives in the
|
``pxar`` is a command line utility for creating and manipulating archives in the
|
||||||
:ref:`pxar-format`.
|
:ref:`pxar-format`.
|
||||||
It is inspired by `casync file archive format
|
It is inspired by `casync file archive format
|
||||||
<http://0pointer.net/blog/casync-a-tool-for-distributing-file-system-images.html>`_,
|
<http://0pointer.net/blog/casync-a-tool-for-distributing-file-system-images.html>`_,
|
||||||
which caters to a similar use-case.
|
which caters to a similar use-case.
|
||||||
The ``.pxar`` format is adapted to fulfill the specific needs of the Proxmox
|
The ``.pxar`` format is adapted to fulfill the specific needs of the Proxmox
|
||||||
Backup Server, for example, efficient storage of hardlinks.
|
Backup Server, for example, efficient storage of hard links.
|
||||||
The format is designed to reduce storage space needed on the server by achieving
|
The format is designed to reduce the required storage on the server by
|
||||||
a high level of deduplication.
|
achieving a high level of deduplication.
|
||||||
|
|
||||||
Creating an Archive
|
Creating an Archive
|
||||||
^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^
|
||||||
@ -24,10 +24,10 @@ This will create a new archive called ``archive.pxar`` with the contents of the
|
|||||||
the same name is already present in the target folder, the creation will
|
the same name is already present in the target folder, the creation will
|
||||||
fail.
|
fail.
|
||||||
|
|
||||||
By default, ``pxar`` will skip certain mountpoints and will not follow device
|
By default, ``pxar`` will skip certain mount points and will not follow device
|
||||||
boundaries. This design decision is based on the primary use case of creating
|
boundaries. This design decision is based on the primary use case of creating
|
||||||
archives for backups. It makes sense to not back up the contents of certain
|
archives for backups. It makes sense to ignore the contents of certain
|
||||||
temporary or system specific files.
|
temporary or system specific files in a backup.
|
||||||
To alter this behavior and follow device boundaries, use the
|
To alter this behavior and follow device boundaries, use the
|
||||||
``--all-file-systems`` flag.
|
``--all-file-systems`` flag.
|
||||||
|
|
||||||
@ -41,40 +41,38 @@ by running:
|
|||||||
|
|
||||||
# pxar create archive.pxar /path/to/source --exclude '**/*.txt'
|
# pxar create archive.pxar /path/to/source --exclude '**/*.txt'
|
||||||
|
|
||||||
Be aware that the shell itself will try to expand all of the glob patterns before
|
Be aware that the shell itself will try to expand glob patterns before invoking
|
||||||
invoking ``pxar``.
|
``pxar``. In order to avoid this, all globs have to be quoted correctly.
|
||||||
In order to avoid this, all globs have to be quoted correctly.
|
|
||||||
|
|
||||||
It is possible to pass the ``--exclude`` parameter multiple times, in order to
|
It is possible to pass the ``--exclude`` parameter multiple times, in order to
|
||||||
match more than one pattern. This allows you to use more complex
|
match more than one pattern. This allows you to use more complex
|
||||||
file exclusion/inclusion behavior. However, it is recommended to use
|
file inclusion/exclusion behavior. However, it is recommended to use
|
||||||
``.pxarexclude`` files instead for such cases.
|
``.pxarexclude`` files instead for such cases.
|
||||||
|
|
||||||
For example you might want to exclude all ``.txt`` files except for a specific
|
For example you might want to exclude all ``.txt`` files except a specific
|
||||||
one from the archive. This is achieved via the negated match pattern, prefixed
|
one from the archive. This would be achieved via the negated match pattern,
|
||||||
by ``!``.
|
prefixed by ``!``. All the glob patterns are relative to the ``source``
|
||||||
All the glob patterns are relative to the ``source`` directory.
|
directory.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
# pxar create archive.pxar /path/to/source --exclude '**/*.txt' --exclude '!/folder/file.txt'
|
# pxar create archive.pxar /path/to/source --exclude '**/*.txt' --exclude '!/folder/file.txt'
|
||||||
|
|
||||||
.. NOTE:: The order of the glob match patterns matters as later ones override
|
.. NOTE:: The order of the glob match patterns matters, as later ones override
|
||||||
previous ones. Permutations of the same patterns lead to different results.
|
earlier ones. Permutations of the same patterns lead to different results.
|
||||||
|
|
||||||
``pxar`` will store the list of glob match patterns passed as parameters via the
|
``pxar`` will store the list of glob match patterns passed as parameters via the
|
||||||
command line, in a file called ``.pxarexclude-cli`` at the root of
|
command line, in a file called ``.pxarexclude-cli``, at the root of the archive.
|
||||||
the archive.
|
|
||||||
If a file with this name is already present in the source folder during archive
|
If a file with this name is already present in the source folder during archive
|
||||||
creation, this file is not included in the archive and the file containing the
|
creation, this file is not included in the archive, and the file containing the
|
||||||
new patterns is added to the archive instead, the original file is not altered.
|
new patterns is added to the archive instead. The original file is not altered.
|
||||||
|
|
||||||
A more convenient and persistent way to exclude files from the archive is by
|
A more convenient and persistent way to exclude files from the archive is by
|
||||||
placing the glob match patterns in ``.pxarexclude`` files.
|
placing the glob match patterns in ``.pxarexclude`` files.
|
||||||
It is possible to create and place these files in any directory of the filesystem
|
It is possible to create and place these files in any directory of the filesystem
|
||||||
tree.
|
tree.
|
||||||
These files must contain one pattern per line, again later patterns win over
|
These files must contain one pattern per line, and later patterns override
|
||||||
previous ones.
|
earlier ones.
|
||||||
The patterns control file exclusions of files present within the given directory
|
The patterns control file exclusions of files present within the given directory
|
||||||
or further below it in the tree.
|
or further below it in the tree.
|
||||||
The behavior is the same as described in :ref:`client_creating_backups`.
|
The behavior is the same as described in :ref:`client_creating_backups`.
|
||||||
@ -89,7 +87,7 @@ with the following command:
|
|||||||
|
|
||||||
# pxar extract archive.pxar /path/to/target
|
# pxar extract archive.pxar /path/to/target
|
||||||
|
|
||||||
If no target is provided, the content of the archive is extracted to the current
|
If no target is provided, the contents of the archive is extracted to the current
|
||||||
working directory.
|
working directory.
|
||||||
|
|
||||||
In order to restore only parts of an archive, single files, and/or folders,
|
In order to restore only parts of an archive, single files, and/or folders,
|
||||||
@ -116,13 +114,13 @@ run the following command:
|
|||||||
# pxar list archive.pxar
|
# pxar list archive.pxar
|
||||||
|
|
||||||
This displays the full path of each file or directory with respect to the
|
This displays the full path of each file or directory with respect to the
|
||||||
archives root.
|
archive's root.
|
||||||
|
|
||||||
Mounting an Archive
|
Mounting an Archive
|
||||||
^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
``pxar`` allows you to mount and inspect the contents of an archive via _`FUSE`.
|
``pxar`` allows you to mount and inspect the contents of an archive via _`FUSE`.
|
||||||
In order to mount an archive named ``archive.pxar`` to the mountpoint ``/mnt``,
|
In order to mount an archive named ``archive.pxar`` to the mount point ``/mnt``,
|
||||||
run the command:
|
run the command:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
@ -130,7 +128,7 @@ run the command:
|
|||||||
# pxar mount archive.pxar /mnt
|
# pxar mount archive.pxar /mnt
|
||||||
|
|
||||||
Once the archive is mounted, you can access its content under the given
|
Once the archive is mounted, you can access its content under the given
|
||||||
mountpoint.
|
mount point.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
|
@ -15,7 +15,7 @@ accessed using the ``disk`` subcommand. This subcommand allows you to initialize
|
|||||||
disks, create various filesystems, and get information about the disks.
|
disks, create various filesystems, and get information about the disks.
|
||||||
|
|
||||||
To view the disks connected to the system, navigate to **Administration ->
|
To view the disks connected to the system, navigate to **Administration ->
|
||||||
Disks** in the web interface or use the ``list`` subcommand of
|
Storage/Disks** in the web interface or use the ``list`` subcommand of
|
||||||
``disk``:
|
``disk``:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
@ -42,9 +42,9 @@ To initialize a disk with a new GPT, use the ``initialize`` subcommand:
|
|||||||
:alt: Create a directory
|
:alt: Create a directory
|
||||||
|
|
||||||
You can create an ``ext4`` or ``xfs`` filesystem on a disk using ``fs
|
You can create an ``ext4`` or ``xfs`` filesystem on a disk using ``fs
|
||||||
create``, or by navigating to **Administration -> Disks -> Directory** in the
|
create``, or by navigating to **Administration -> Storage/Disks -> Directory**
|
||||||
web interface and creating one from there. The following command creates an
|
in the web interface and creating one from there. The following command creates
|
||||||
``ext4`` filesystem and passes the ``--add-datastore`` parameter, in order to
|
an ``ext4`` filesystem and passes the ``--add-datastore`` parameter, in order to
|
||||||
automatically create a datastore on the disk (in this case ``sdd``). This will
|
automatically create a datastore on the disk (in this case ``sdd``). This will
|
||||||
create a datastore at the location ``/mnt/datastore/store1``:
|
create a datastore at the location ``/mnt/datastore/store1``:
|
||||||
|
|
||||||
@ -57,7 +57,7 @@ create a datastore at the location ``/mnt/datastore/store1``:
|
|||||||
:alt: Create ZFS
|
:alt: Create ZFS
|
||||||
|
|
||||||
You can also create a ``zpool`` with various raid levels from **Administration
|
You can also create a ``zpool`` with various raid levels from **Administration
|
||||||
-> Disks -> Zpool** in the web interface, or by using ``zpool create``. The command
|
-> Storage/Disks -> ZFS** in the web interface, or by using ``zpool create``. The command
|
||||||
below creates a mirrored ``zpool`` using two disks (``sdb`` & ``sdc``) and
|
below creates a mirrored ``zpool`` using two disks (``sdb`` & ``sdc``) and
|
||||||
mounts it under ``/mnt/datastore/zpool1``:
|
mounts it under ``/mnt/datastore/zpool1``:
|
||||||
|
|
||||||
@ -102,7 +102,7 @@ is stored in the file ``/etc/proxmox-backup/datastore.cfg``.
|
|||||||
subdirectories per directory. That number comes from the 2\ :sup:`16`
|
subdirectories per directory. That number comes from the 2\ :sup:`16`
|
||||||
pre-created chunk namespace directories, and the ``.`` and ``..`` default
|
pre-created chunk namespace directories, and the ``.`` and ``..`` default
|
||||||
directory entries. This requirement excludes certain filesystems and
|
directory entries. This requirement excludes certain filesystems and
|
||||||
filesystem configuration from being supported for a datastore. For example,
|
filesystem configurations from being supported for a datastore. For example,
|
||||||
``ext3`` as a whole or ``ext4`` with the ``dir_nlink`` feature manually disabled.
|
``ext3`` as a whole or ``ext4`` with the ``dir_nlink`` feature manually disabled.
|
||||||
|
|
||||||
|
|
||||||
@ -113,14 +113,15 @@ Datastore Configuration
|
|||||||
:align: right
|
:align: right
|
||||||
:alt: Datastore Overview
|
:alt: Datastore Overview
|
||||||
|
|
||||||
You can configure multiple datastores. Minimum one datastore needs to be
|
You can configure multiple datastores. A minimum of one datastore needs to be
|
||||||
configured. The datastore is identified by a simple *name* and points to a
|
configured. The datastore is identified by a simple *name* and points to a
|
||||||
directory on the filesystem. Each datastore also has associated retention
|
directory on the filesystem. Each datastore also has associated retention
|
||||||
settings of how many backup snapshots for each interval of ``hourly``,
|
settings of how many backup snapshots for each interval of ``hourly``,
|
||||||
``daily``, ``weekly``, ``monthly``, ``yearly`` as well as a time-independent
|
``daily``, ``weekly``, ``monthly``, ``yearly`` as well as a time-independent
|
||||||
number of backups to keep in that store. :ref:`backup-pruning` and
|
number of backups to keep in that store. :ref:`backup-pruning` and
|
||||||
:ref:`garbage collection <client_garbage-collection>` can also be configured to run
|
:ref:`garbage collection <client_garbage-collection>` can also be configured to
|
||||||
periodically based on a configured schedule (see :ref:`calendar-event-scheduling`) per datastore.
|
run periodically, based on a configured schedule (see
|
||||||
|
:ref:`calendar-event-scheduling`) per datastore.
|
||||||
|
|
||||||
|
|
||||||
.. _storage_datastore_create:
|
.. _storage_datastore_create:
|
||||||
@ -146,7 +147,8 @@ window:
|
|||||||
* *Comment* can be used to add some contextual information to the datastore.
|
* *Comment* can be used to add some contextual information to the datastore.
|
||||||
|
|
||||||
Alternatively you can create a new datastore from the command line. The
|
Alternatively you can create a new datastore from the command line. The
|
||||||
following command creates a new datastore called ``store1`` on :file:`/backup/disk1/store1`
|
following command creates a new datastore called ``store1`` on
|
||||||
|
:file:`/backup/disk1/store1`
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -156,7 +158,7 @@ following command creates a new datastore called ``store1`` on :file:`/backup/di
|
|||||||
Managing Datastores
|
Managing Datastores
|
||||||
^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
To list existing datastores from the command line run:
|
To list existing datastores from the command line, run:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -216,8 +218,9 @@ After creating a datastore, the following default layout will appear:
|
|||||||
|
|
||||||
`.lock` is an empty file used for process locking.
|
`.lock` is an empty file used for process locking.
|
||||||
|
|
||||||
The `.chunks` directory contains folders, starting from `0000` and taking hexadecimal values until `ffff`. These
|
The `.chunks` directory contains folders, starting from `0000` and increasing in
|
||||||
directories will store the chunked data after a backup operation has been executed.
|
hexadecimal values until `ffff`. These directories will store the chunked data,
|
||||||
|
categorized by checksum, after a backup operation has been executed.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
|
@ -4,8 +4,8 @@ Host System Administration
|
|||||||
==========================
|
==========================
|
||||||
|
|
||||||
`Proxmox Backup`_ is based on the famous Debian_ Linux
|
`Proxmox Backup`_ is based on the famous Debian_ Linux
|
||||||
distribution. That means that you have access to the whole world of
|
distribution. This means that you have access to the entire range of
|
||||||
Debian packages, and the base system is well documented. The `Debian
|
Debian packages, and that the base system is well documented. The `Debian
|
||||||
Administrator's Handbook`_ is available online, and provides a
|
Administrator's Handbook`_ is available online, and provides a
|
||||||
comprehensive introduction to the Debian operating system.
|
comprehensive introduction to the Debian operating system.
|
||||||
|
|
||||||
@ -17,11 +17,11 @@ updates to some Debian packages when necessary.
|
|||||||
|
|
||||||
We also deliver a specially optimized Linux kernel, where we enable
|
We also deliver a specially optimized Linux kernel, where we enable
|
||||||
all required virtualization and container features. That kernel
|
all required virtualization and container features. That kernel
|
||||||
includes drivers for ZFS_, and several hardware drivers. For example,
|
includes drivers for ZFS_, as well as several hardware drivers. For example,
|
||||||
we ship Intel network card drivers to support their newest hardware.
|
we ship Intel network card drivers to support their newest hardware.
|
||||||
|
|
||||||
The following sections will concentrate on backup related topics. They
|
The following sections will concentrate on backup related topics. They
|
||||||
either explain things which are different on `Proxmox Backup`_, or
|
will explain things which are different on `Proxmox Backup`_, or
|
||||||
tasks which are commonly used on `Proxmox Backup`_. For other topics,
|
tasks which are commonly used on `Proxmox Backup`_. For other topics,
|
||||||
please refer to the standard Debian documentation.
|
please refer to the standard Debian documentation.
|
||||||
|
|
||||||
|
@ -3,9 +3,6 @@
|
|||||||
Tape Backup
|
Tape Backup
|
||||||
===========
|
===========
|
||||||
|
|
||||||
.. CAUTION:: Tape Backup is a technical preview feature, not meant for
|
|
||||||
production use.
|
|
||||||
|
|
||||||
.. image:: images/screenshots/pbs-gui-tape-changer-overview.png
|
.. image:: images/screenshots/pbs-gui-tape-changer-overview.png
|
||||||
:align: right
|
:align: right
|
||||||
:alt: Tape Backup: Tape changer overview
|
:alt: Tape Backup: Tape changer overview
|
||||||
@ -67,8 +64,10 @@ tape compression feature has no advantage.
|
|||||||
Supported Hardware
|
Supported Hardware
|
||||||
------------------
|
------------------
|
||||||
|
|
||||||
Proxmox Backup Server supports `Linear Tape-Open`_ generation 4 (LTO-4)
|
Proxmox Backup Server supports `Linear Tape-Open`_ generation 5 (LTO-5)
|
||||||
or later.
|
or later and has best-effort support for generation 4 (LTO-4). While
|
||||||
|
many LTO-4 systems are known to work, some might need firmware updates or
|
||||||
|
do not implement necessary features to work with Proxmox Backup Server.
|
||||||
|
|
||||||
Tape changing is carried out using the SCSI Medium Changer protocol,
|
Tape changing is carried out using the SCSI Medium Changer protocol,
|
||||||
so all modern tape libraries should work.
|
so all modern tape libraries should work.
|
||||||
@ -846,6 +845,17 @@ Update Inventory
|
|||||||
Restore Catalog
|
Restore Catalog
|
||||||
~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
|
To restore a catalog from an existing tape, just insert the tape into the drive
|
||||||
|
and execute:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-tape catalog
|
||||||
|
|
||||||
|
|
||||||
|
You can restore from a tape even without an existing catalog, but only the
|
||||||
|
whole media set. If you do this, the catalog will be automatically created.
|
||||||
|
|
||||||
|
|
||||||
Encryption Key Management
|
Encryption Key Management
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
@ -8,7 +8,7 @@ Datastores
|
|||||||
|
|
||||||
A Datastore is the logical place where :ref:`Backup Snapshots
|
A Datastore is the logical place where :ref:`Backup Snapshots
|
||||||
<term_backup_snapshot>` and their chunks are stored. Snapshots consist of a
|
<term_backup_snapshot>` and their chunks are stored. Snapshots consist of a
|
||||||
manifest, blobs, dynamic- and fixed-indexes (see :ref:`terms`), and are
|
manifest, blobs, and dynamic- and fixed-indexes (see :ref:`terms`), and are
|
||||||
stored in the following directory structure:
|
stored in the following directory structure:
|
||||||
|
|
||||||
<datastore-root>/<type>/<id>/<time>/
|
<datastore-root>/<type>/<id>/<time>/
|
||||||
@ -32,8 +32,8 @@ The chunks of a datastore are found in
|
|||||||
|
|
||||||
<datastore-root>/.chunks/
|
<datastore-root>/.chunks/
|
||||||
|
|
||||||
This chunk directory is further subdivided by the first four byte of the chunks
|
This chunk directory is further subdivided by the first four bytes of the
|
||||||
checksum, so the chunk with the checksum
|
chunk's checksum, so a chunk with the checksum
|
||||||
|
|
||||||
a342e8151cbf439ce65f3df696b54c67a114982cc0aa751f2852c2f7acc19a8b
|
a342e8151cbf439ce65f3df696b54c67a114982cc0aa751f2852c2f7acc19a8b
|
||||||
|
|
||||||
@ -47,7 +47,7 @@ per directory can be bad for file system performance.
|
|||||||
These chunk directories ('0000'-'ffff') will be preallocated when a datastore
|
These chunk directories ('0000'-'ffff') will be preallocated when a datastore
|
||||||
is created.
|
is created.
|
||||||
|
|
||||||
Fixed-sized Chunks
|
Fixed-Sized Chunks
|
||||||
^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
For block based backups (like VMs), fixed-sized chunks are used. The content
|
For block based backups (like VMs), fixed-sized chunks are used. The content
|
||||||
@ -58,10 +58,10 @@ often tries to allocate files in contiguous pieces, so new files get new
|
|||||||
blocks, and changing existing files changes only their own blocks.
|
blocks, and changing existing files changes only their own blocks.
|
||||||
|
|
||||||
As an optimization, VMs in `Proxmox VE`_ can make use of 'dirty bitmaps', which
|
As an optimization, VMs in `Proxmox VE`_ can make use of 'dirty bitmaps', which
|
||||||
can track the changed blocks of an image. Since these bitmap are also a
|
can track the changed blocks of an image. Since these bitmaps are also a
|
||||||
representation of the image split into chunks, there is a direct relation
|
representation of the image split into chunks, there is a direct relation
|
||||||
between dirty blocks of the image and chunks which need to get uploaded, so
|
between the dirty blocks of the image and chunks which need to be uploaded.
|
||||||
only modified chunks of the disk have to be uploaded for a backup.
|
Thus, only modified chunks of the disk need to be uploaded to a backup.
|
||||||
|
|
||||||
Since the image is always split into chunks of the same size, unchanged blocks
|
Since the image is always split into chunks of the same size, unchanged blocks
|
||||||
will result in identical checksums for those chunks, so such chunks do not need
|
will result in identical checksums for those chunks, so such chunks do not need
|
||||||
@ -71,24 +71,24 @@ changed blocks.
|
|||||||
For consistency, `Proxmox VE`_ uses a QEMU internal snapshot mechanism, that
|
For consistency, `Proxmox VE`_ uses a QEMU internal snapshot mechanism, that
|
||||||
does not rely on storage snapshots either.
|
does not rely on storage snapshots either.
|
||||||
|
|
||||||
Dynamically sized Chunks
|
Dynamically Sized Chunks
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
If one does not want to backup block-based systems but rather file-based
|
When working with file-based systems rather than block-based systems,
|
||||||
systems, using fixed-sized chunks is not a good idea, since every time a file
|
using fixed-sized chunks is not a good idea, since every time a file
|
||||||
would change in size, the remaining data gets shifted around and this would
|
would change in size, the remaining data would be shifted around,
|
||||||
result in many chunks changing, reducing the amount of deduplication.
|
resulting in many chunks changing and the amount of deduplication being reduced.
|
||||||
|
|
||||||
To improve this, `Proxmox Backup`_ Server uses dynamically sized chunks
|
To improve this, `Proxmox Backup`_ Server uses dynamically sized chunks
|
||||||
instead. Instead of splitting an image into fixed sizes, it first generates a
|
instead. Instead of splitting an image into fixed sizes, it first generates a
|
||||||
consistent file archive (:ref:`pxar <pxar-format>`) and uses a rolling hash
|
consistent file archive (:ref:`pxar <pxar-format>`) and uses a rolling hash
|
||||||
over this on-the-fly generated archive to calculate chunk boundaries.
|
over this on-the-fly generated archive to calculate chunk boundaries.
|
||||||
|
|
||||||
We use a variant of Buzhash which is a cyclic polynomial algorithm. It works
|
We use a variant of Buzhash which is a cyclic polynomial algorithm. It works
|
||||||
by continuously calculating a checksum while iterating over the data, and on
|
by continuously calculating a checksum while iterating over the data, and on
|
||||||
certain conditions it triggers a hash boundary.
|
certain conditions, it triggers a hash boundary.
|
||||||
|
|
||||||
Assuming that most files of the system that is to be backed up have not
|
Assuming that most files on the system that is to be backed up have not
|
||||||
changed, eventually the algorithm triggers the boundary on the same data as a
|
changed, eventually the algorithm triggers the boundary on the same data as a
|
||||||
previous backup, resulting in chunks that can be reused.
|
previous backup, resulting in chunks that can be reused.
|
||||||
|
|
||||||
@ -100,8 +100,8 @@ can be encrypted, and they are handled in a slightly different manner than
|
|||||||
normal chunks.
|
normal chunks.
|
||||||
|
|
||||||
The hashes of encrypted chunks are calculated not with the actual (encrypted)
|
The hashes of encrypted chunks are calculated not with the actual (encrypted)
|
||||||
chunk content, but with the plain-text content concatenated with the encryption
|
chunk content, but with the plain-text content, concatenated with the encryption
|
||||||
key. This way, two chunks of the same data encrypted with different keys
|
key. This way, two chunks with the same data but encrypted with different keys
|
||||||
generate two different checksums and no collisions occur for multiple
|
generate two different checksums and no collisions occur for multiple
|
||||||
encryption keys.
|
encryption keys.
|
||||||
|
|
||||||
@ -112,14 +112,14 @@ the previous backup, do not need to be encrypted and uploaded.
|
|||||||
Caveats and Limitations
|
Caveats and Limitations
|
||||||
-----------------------
|
-----------------------
|
||||||
|
|
||||||
Notes on hash collisions
|
Notes on Hash Collisions
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
Every hashing algorithm has a chance to produce collisions, meaning two (or
|
Every hashing algorithm has a chance to produce collisions, meaning two (or
|
||||||
more) inputs generate the same checksum. For SHA-256, this chance is
|
more) inputs generate the same checksum. For SHA-256, this chance is
|
||||||
negligible. To calculate such a collision, one can use the ideas of the
|
negligible. To calculate the chances of such a collision, one can use the ideas
|
||||||
'birthday problem' from probability theory. For big numbers, this is actually
|
of the 'birthday problem' from probability theory. For big numbers, this is
|
||||||
infeasible to calculate with regular computers, but there is a good
|
actually unfeasible to calculate with regular computers, but there is a good
|
||||||
approximation:
|
approximation:
|
||||||
|
|
||||||
.. math::
|
.. math::
|
||||||
@ -127,7 +127,7 @@ approximation:
|
|||||||
p(n, d) = 1 - e^{-n^2/(2d)}
|
p(n, d) = 1 - e^{-n^2/(2d)}
|
||||||
|
|
||||||
Where `n` is the number of tries, and `d` is the number of possibilities.
|
Where `n` is the number of tries, and `d` is the number of possibilities.
|
||||||
For a concrete example lets assume a large datastore of 1 PiB, and an average
|
For a concrete example, lets assume a large datastore of 1 PiB and an average
|
||||||
chunk size of 4 MiB. That means :math:`n = 268435456` tries, and :math:`d =
|
chunk size of 4 MiB. That means :math:`n = 268435456` tries, and :math:`d =
|
||||||
2^{256}` possibilities. Inserting those values in the formula from earlier you
|
2^{256}` possibilities. Inserting those values in the formula from earlier you
|
||||||
will see that the probability of a collision in that scenario is:
|
will see that the probability of a collision in that scenario is:
|
||||||
@ -136,31 +136,96 @@ will see that the probability of a collision in that scenario is:
|
|||||||
|
|
||||||
3.1115 * 10^{-61}
|
3.1115 * 10^{-61}
|
||||||
|
|
||||||
For context, in a lottery game of guessing 6 out of 45, the chance to correctly
|
For context, in a lottery game of guessing 6 numbers out of 45, the chance to
|
||||||
guess all 6 numbers is only :math:`1.2277 * 10^{-7}`, that means the chance of
|
correctly guess all 6 numbers is only :math:`1.2277 * 10^{-7}`. This means the
|
||||||
a collision is about the same as winning 13 such lotto games *in a row*.
|
chance of a collision is about the same as winning 13 such lottery games *in a
|
||||||
|
row*.
|
||||||
|
|
||||||
In conclusion, it is extremely unlikely that such a collision would occur by
|
In conclusion, it is extremely unlikely that such a collision would occur by
|
||||||
accident in a normal datastore.
|
accident in a normal datastore.
|
||||||
|
|
||||||
Additionally, SHA-256 is prone to length extension attacks, but since there is
|
Additionally, SHA-256 is prone to length extension attacks, but since there is
|
||||||
an upper limit for how big the chunk are, this is not a problem, since a
|
an upper limit for how big the chunks are, this is not a problem, because a
|
||||||
potential attacker cannot arbitrarily add content to the data beyond that
|
potential attacker cannot arbitrarily add content to the data beyond that
|
||||||
limit.
|
limit.
|
||||||
|
|
||||||
File-based Backup
|
File-Based Backup
|
||||||
^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
Since dynamically sized chunks (for file-based backups) are created on a custom
|
Since dynamically sized chunks (for file-based backups) are created on a custom
|
||||||
archive format (pxar) and not over the files directly, there is no relation
|
archive format (pxar) and not over the files directly, there is no relation
|
||||||
between files and the chunks. This means that the Proxmox Backup client has to
|
between the files and chunks. This means that the Proxmox Backup Client has to
|
||||||
read all files again for every backup, otherwise it would not be possible to
|
read all files again for every backup, otherwise it would not be possible to
|
||||||
generate a consistent independent pxar archive where the original chunks can be
|
generate a consistent, independent pxar archive where the original chunks can be
|
||||||
reused. Note that there will be still only new or change chunks be uploaded.
|
reused. Note that in spite of this, only new or changed chunks will be uploaded.
|
||||||
|
|
||||||
Verification of encrypted chunks
|
Verification of Encrypted Chunks
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
For encrypted chunks, only the checksum of the original (plaintext) data is
|
For encrypted chunks, only the checksum of the original (plaintext) data is
|
||||||
available, making it impossible for the server (without the encryption key), to
|
available, making it impossible for the server (without the encryption key) to
|
||||||
verify its content against it. Instead only the CRC-32 checksum gets checked.
|
verify its content against it. Instead only the CRC-32 checksum gets checked.
|
||||||
|
|
||||||
|
Troubleshooting
|
||||||
|
---------------
|
||||||
|
|
||||||
|
Index files(*.fidx*, *.didx*) contain information about how to rebuild a file.
|
||||||
|
More precisely, they contain an ordered list of references to the chunks that
|
||||||
|
the original file was split into. If there is something wrong with a snapshot,
|
||||||
|
it might be useful to find out which chunks are referenced in it, and check
|
||||||
|
whether they are present and intact. The ``proxmox-backup-debug`` command line
|
||||||
|
tool can be used to inspect such files and recover their contents. For example,
|
||||||
|
to get a list of the referenced chunks of a *.fidx* index:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-debug inspect file drive-scsi0.img.fidx
|
||||||
|
|
||||||
|
The same command can be used to inspect *.blob* files. Without the ``--decode``
|
||||||
|
parameter, just the size and the encryption type, if any, are printed. If
|
||||||
|
``--decode`` is set, the blob file is decoded into the specified file ('-' will
|
||||||
|
decode it directly to stdout).
|
||||||
|
|
||||||
|
The following example would print the decoded contents of
|
||||||
|
`qemu-server.conf.blob`. If the file you're trying to inspect is encrypted, a
|
||||||
|
path to the key file must be provided using ``--keyfile``.
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-debug inspect file qemu-server.conf.blob --decode -
|
||||||
|
|
||||||
|
You can also check in which index files a specific chunk file is referenced
|
||||||
|
with:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-debug inspect chunk b531d3ffc9bd7c65748a61198c060678326a431db7eded874c327b7986e595e0 --reference-filter /path/in/a/datastore/directory
|
||||||
|
|
||||||
|
Here ``--reference-filter`` specifies where index files should be searched. This
|
||||||
|
can be an arbitrary path. If, for some reason, the filename of the chunk was
|
||||||
|
changed, you can explicitly specify the digest using ``--digest``. By default, the
|
||||||
|
chunk filename is used as the digest to look for. If no ``--reference-filter``
|
||||||
|
is specified, it will only print the CRC and encryption status of the chunk. You
|
||||||
|
can also decode chunks, by setting the ``--decode`` flag. If the chunk is
|
||||||
|
encrypted, a ``--keyfile`` must be provided, in order to decode it.
|
||||||
|
|
||||||
|
Restore without a Running Proxmox Backup Server
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
|
It's possible to restore specific files from a snapshot, without a running
|
||||||
|
Proxmox Backup Server instance, using the ``recover`` subcommand, provided you
|
||||||
|
have access to the intact index and chunk files. Note that you also need the
|
||||||
|
corresponding key file if the backup was encrypted.
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-debug recover index drive-scsi0.img.fidx /path/to/.chunks
|
||||||
|
|
||||||
|
In the above example, the `/path/to/.chunks` argument is the path to the
|
||||||
|
directory that contains the chunks, and `drive-scsi0.img.fidx` is the index file
|
||||||
|
of the file you'd like to restore. Both paths can be absolute or relative. With
|
||||||
|
``--skip-crc``, it's possible to disable the CRC checks of the chunks. This
|
||||||
|
will speed up the process slightly and allow for trying to restore (partially)
|
||||||
|
corrupt chunks. It's recommended to always try without the skip-CRC option
|
||||||
|
first.
|
||||||
|
|
||||||
|
@ -41,23 +41,23 @@ Binary Data (BLOBs)
|
|||||||
~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
This type is used to store smaller (< 16MB) binary data such as
|
This type is used to store smaller (< 16MB) binary data such as
|
||||||
configuration files. Larger files should be stored as image archive.
|
configuration files. Larger files should be stored as image archives.
|
||||||
|
|
||||||
.. caution:: Please do not store all files as BLOBs. Instead, use the
|
.. caution:: Please do not store all files as BLOBs. Instead, use the
|
||||||
file archive to store whole directory trees.
|
file archive to store entire directory trees.
|
||||||
|
|
||||||
|
|
||||||
Catalog File: ``catalog.pcat1``
|
Catalog File: ``catalog.pcat1``
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
The catalog file is an index for file archives. It contains
|
The catalog file is an index for file archives. It contains
|
||||||
the list of files and is used to speed up search operations.
|
the list of included files and is used to speed up search operations.
|
||||||
|
|
||||||
|
|
||||||
The Manifest: ``index.json``
|
The Manifest: ``index.json``
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
The manifest contains the list of all backup files, their
|
The manifest contains a list of all backed up files, and their
|
||||||
sizes and checksums. It is used to verify the consistency of a
|
sizes and checksums. It is used to verify the consistency of a
|
||||||
backup.
|
backup.
|
||||||
|
|
||||||
@ -68,18 +68,19 @@ Backup Type
|
|||||||
The backup server groups backups by *type*, where *type* is one of:
|
The backup server groups backups by *type*, where *type* is one of:
|
||||||
|
|
||||||
``vm``
|
``vm``
|
||||||
This type is used for :term:`virtual machine`\ s. Typically
|
This type is used for :term:`virtual machine`\ s. It typically
|
||||||
consists of the virtual machine's configuration file and an image archive
|
consists of the virtual machine's configuration file and an image archive
|
||||||
for each disk.
|
for each disk.
|
||||||
|
|
||||||
``ct``
|
``ct``
|
||||||
This type is used for :term:`container`\ s. Consists of the container's
|
This type is used for :term:`container`\ s. It consists of the container's
|
||||||
configuration and a single file archive for the filesystem content.
|
configuration and a single file archive for the filesystem's contents.
|
||||||
|
|
||||||
``host``
|
``host``
|
||||||
This type is used for backups created from within the backed up machine.
|
This type is used for file/directory backups created from within a machine.
|
||||||
Typically this would be a physical host but could also be a virtual machine
|
Typically this would be a physical host, but could also be a virtual machine
|
||||||
or container. Such backups may contain file and image archives, there are no restrictions in this regard.
|
or container. Such backups may contain file and image archives; there are no
|
||||||
|
restrictions in this regard.
|
||||||
|
|
||||||
|
|
||||||
Backup ID
|
Backup ID
|
||||||
|
101
docs/traffic-control.rst
Normal file
101
docs/traffic-control.rst
Normal file
@ -0,0 +1,101 @@
|
|||||||
|
.. _sysadmin_traffic_control:
|
||||||
|
|
||||||
|
Traffic Control
|
||||||
|
---------------
|
||||||
|
|
||||||
|
.. image:: images/screenshots/pbs-gui-traffic-control-add.png
|
||||||
|
:align: right
|
||||||
|
:alt: Add a traffic control limit
|
||||||
|
|
||||||
|
Creating and restoring backups can produce lots of traffic and impact other
|
||||||
|
users of the network or shared storages.
|
||||||
|
|
||||||
|
Proxmox Backup Server allows to limit network traffic for clients within
|
||||||
|
specified networks using a token bucket filter (TBF).
|
||||||
|
|
||||||
|
This allows you to avoid network congestion or to prioritize traffic from
|
||||||
|
certain hosts.
|
||||||
|
|
||||||
|
You can manage the traffic controls either over the web-interface or using the
|
||||||
|
``traffic-control`` commandos of the ``proxmox-backup-manager`` command-line
|
||||||
|
tool.
|
||||||
|
|
||||||
|
.. note:: Sync jobs on the server are not affected by its rate-in limits. If
|
||||||
|
you want to limit the incomming traffic that a pull-based sync job
|
||||||
|
generates, you need to setup a job-specific rate-in limit. See
|
||||||
|
:ref:`syncjobs`.
|
||||||
|
|
||||||
|
The following command adds a traffic control rule to limit all IPv4 clients
|
||||||
|
(network ``0.0.0.0/0``) to 100 MB/s:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-manager traffic-control create rule0 --network 0.0.0.0/0 \
|
||||||
|
--rate-in 100MB --rate-out 100MB \
|
||||||
|
--comment "Default rate limit (100MB/s) for all clients"
|
||||||
|
|
||||||
|
.. note:: To limit both IPv4 and IPv6 network spaces you need to pass two
|
||||||
|
network parameters ``::/0`` and ``0.0.0.0/0``.
|
||||||
|
|
||||||
|
It is possible to restrict rules to certain time frames, for example the
|
||||||
|
company office hours:
|
||||||
|
|
||||||
|
.. tip:: You can use SI (base 10: KB, MB, ...) or IEC (base 2: KiB, MiB, ...)
|
||||||
|
units.
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-manager traffic-control update rule0 \
|
||||||
|
--timeframe "mon..fri 8-12" \
|
||||||
|
--timeframe "mon..fri 14:30-18"
|
||||||
|
|
||||||
|
If there are more rules, the server uses the rule with the smaller network. For
|
||||||
|
example, we can overwrite the setting for our private network (and the server
|
||||||
|
itself) with:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-manager traffic-control create rule1 \
|
||||||
|
--network 192.168.2.0/24 \
|
||||||
|
--network 127.0.0.0/8 \
|
||||||
|
--rate-in 20GB --rate-out 20GB \
|
||||||
|
--comment "Use 20GB/s for the local network"
|
||||||
|
|
||||||
|
.. note:: The behavior is undefined if there are several rules for the same network.
|
||||||
|
|
||||||
|
If there are multiple rules that match the same network all of them will be
|
||||||
|
applied, which means that the smallest one wins, as it's bucket fills up the
|
||||||
|
fastest.
|
||||||
|
|
||||||
|
To list the current rules use:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-manager traffic-control list
|
||||||
|
┌───────┬─────────────┬─────────────┬─────────────────────────┬────────────...─┐
|
||||||
|
│ name │ rate-in │ rate-out │ network │ timeframe ... │
|
||||||
|
╞═══════╪═════════════╪═════════════╪═════════════════════════╪════════════...═╡
|
||||||
|
│ rule0 │ 100 MB │ 100 MB │ ["0.0.0.0/0"] │ ["mon..fri ... │
|
||||||
|
├───────┼─────────────┼─────────────┼─────────────────────────┼────────────...─┤
|
||||||
|
│ rule1 │ 20 GB │ 20 GB │ ["192.168.2.0/24", ...] │ ... │
|
||||||
|
└───────┴─────────────┴─────────────┴─────────────────────────┴────────────...─┘
|
||||||
|
|
||||||
|
Rules can also be removed:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-manager traffic-control remove rule1
|
||||||
|
|
||||||
|
|
||||||
|
To show the state (current data rate) of all configured rules use:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
# proxmox-backup-manager traffic-control traffic
|
||||||
|
┌───────┬─────────────┬──────────────┐
|
||||||
|
│ name │ cur-rate-in │ cur-rate-out │
|
||||||
|
╞═══════╪═════════════╪══════════════╡
|
||||||
|
│ rule0 │ 0 B │ 0 B │
|
||||||
|
├───────┼─────────────┼──────────────┤
|
||||||
|
│ rule1 │ 1.161 GiB │ 19.146 KiB │
|
||||||
|
└───────┴─────────────┴──────────────┘
|
@ -15,17 +15,19 @@ Proxmox Backup Server supports several authentication realms, and you need to
|
|||||||
choose the realm when you add a new user. Possible realms are:
|
choose the realm when you add a new user. Possible realms are:
|
||||||
|
|
||||||
:pam: Linux PAM standard authentication. Use this if you want to
|
:pam: Linux PAM standard authentication. Use this if you want to
|
||||||
authenticate as Linux system user (Users need to exist on the
|
authenticate as a Linux system user (users need to exist on the
|
||||||
system).
|
system).
|
||||||
|
|
||||||
:pbs: Proxmox Backup Server realm. This type stores hashed passwords in
|
:pbs: Proxmox Backup Server realm. This type stores hashed passwords in
|
||||||
``/etc/proxmox-backup/shadow.json``.
|
``/etc/proxmox-backup/shadow.json``.
|
||||||
|
|
||||||
After installation, there is a single user ``root@pam``, which
|
:openid: OpenID Connect server. Users can authenticate against an external
|
||||||
corresponds to the Unix superuser. User configuration information is stored in the file
|
OpenID Connect server.
|
||||||
``/etc/proxmox-backup/user.cfg``. You can use the
|
|
||||||
``proxmox-backup-manager`` command line tool to list or manipulate
|
After installation, there is a single user, ``root@pam``, which corresponds to
|
||||||
users:
|
the Unix superuser. User configuration information is stored in the file
|
||||||
|
``/etc/proxmox-backup/user.cfg``. You can use the ``proxmox-backup-manager``
|
||||||
|
command line tool to list or manipulate users:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -40,13 +42,13 @@ users:
|
|||||||
:align: right
|
:align: right
|
||||||
:alt: Add a new user
|
:alt: Add a new user
|
||||||
|
|
||||||
The superuser has full administration rights on everything, so you
|
The superuser has full administration rights on everything, so it's recommended
|
||||||
normally want to add other users with less privileges. You can add a new
|
to add other users with less privileges. You can add a new
|
||||||
user with the ``user create`` subcommand or through the web
|
user with the ``user create`` subcommand or through the web
|
||||||
interface, under the **User Management** tab of **Configuration -> Access
|
interface, under the **User Management** tab of **Configuration -> Access
|
||||||
Control**. The ``create`` subcommand lets you specify many options like
|
Control**. The ``create`` subcommand lets you specify many options like
|
||||||
``--email`` or ``--password``. You can update or change any user properties
|
``--email`` or ``--password``. You can update or change any user properties
|
||||||
using the ``update`` subcommand later (**Edit** in the GUI):
|
using the ``user update`` subcommand later (**Edit** in the GUI):
|
||||||
|
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
@ -71,16 +73,16 @@ The resulting user list looks like this:
|
|||||||
│ root@pam │ 1 │ │ │ │ │ Superuser │
|
│ root@pam │ 1 │ │ │ │ │ Superuser │
|
||||||
└──────────┴────────┴────────┴───────────┴──────────┴──────────────────┴──────────────────┘
|
└──────────┴────────┴────────┴───────────┴──────────┴──────────────────┴──────────────────┘
|
||||||
|
|
||||||
Newly created users do not have any permissions. Please read the Access Control
|
Newly created users do not have any permissions. Please read the :ref:`user_acl`
|
||||||
section to learn how to set access permissions.
|
section to learn how to set access permissions.
|
||||||
|
|
||||||
If you want to disable a user account, you can do that by setting ``--enable`` to ``0``
|
You can disable a user account by setting ``--enable`` to ``0``:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
# proxmox-backup-manager user update john@pbs --enable 0
|
# proxmox-backup-manager user update john@pbs --enable 0
|
||||||
|
|
||||||
Or completely remove the user with:
|
Or completely remove a user with:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -95,7 +97,7 @@ API Tokens
|
|||||||
:align: right
|
:align: right
|
||||||
:alt: API Token Overview
|
:alt: API Token Overview
|
||||||
|
|
||||||
Any authenticated user can generate API tokens which can in turn be used to
|
Any authenticated user can generate API tokens, which can in turn be used to
|
||||||
configure various clients, instead of directly providing the username and
|
configure various clients, instead of directly providing the username and
|
||||||
password.
|
password.
|
||||||
|
|
||||||
@ -117,7 +119,7 @@ The API token is passed from the client to the server by setting the
|
|||||||
``Authorization`` HTTP header with method ``PBSAPIToken`` to the value
|
``Authorization`` HTTP header with method ``PBSAPIToken`` to the value
|
||||||
``TOKENID:TOKENSECRET``.
|
``TOKENID:TOKENSECRET``.
|
||||||
|
|
||||||
Generating new tokens can done using ``proxmox-backup-manager`` or the GUI:
|
You can generate tokens from the GUI or by using ``proxmox-backup-manager``:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@ -154,9 +156,9 @@ section to learn how to set access permissions.
|
|||||||
Access Control
|
Access Control
|
||||||
--------------
|
--------------
|
||||||
|
|
||||||
By default new users and API tokens do not have any permission. Instead you
|
By default, new users and API tokens do not have any permissions. Instead you
|
||||||
need to specify what is allowed and what is not. You can do this by assigning
|
need to specify what is allowed and what is not. You can do this by assigning
|
||||||
roles to users/tokens on specific objects like datastores or remotes. The
|
roles to users/tokens on specific objects, like datastores or remotes. The
|
||||||
following roles exist:
|
following roles exist:
|
||||||
|
|
||||||
**NoAccess**
|
**NoAccess**
|
||||||
@ -176,7 +178,7 @@ following roles exist:
|
|||||||
is not allowed to read the actual data.
|
is not allowed to read the actual data.
|
||||||
|
|
||||||
**DatastoreReader**
|
**DatastoreReader**
|
||||||
Can Inspect datastore content and can do restores.
|
Can Inspect datastore content and do restores.
|
||||||
|
|
||||||
**DatastoreBackup**
|
**DatastoreBackup**
|
||||||
Can backup and restore owned backups.
|
Can backup and restore owned backups.
|
||||||
@ -193,6 +195,18 @@ following roles exist:
|
|||||||
**RemoteSyncOperator**
|
**RemoteSyncOperator**
|
||||||
Is allowed to read data from a remote.
|
Is allowed to read data from a remote.
|
||||||
|
|
||||||
|
**TapeAudit**
|
||||||
|
Can view tape related configuration and status
|
||||||
|
|
||||||
|
**TapeAdministrat**
|
||||||
|
Can do anything related to tape backup
|
||||||
|
|
||||||
|
**TapeOperator**
|
||||||
|
Can do tape backup and restore (but no configuration changes)
|
||||||
|
|
||||||
|
**TapeReader**
|
||||||
|
Can read and inspect tape configuration and media content
|
||||||
|
|
||||||
.. image:: images/screenshots/pbs-gui-user-management-add-user.png
|
.. image:: images/screenshots/pbs-gui-user-management-add-user.png
|
||||||
:align: right
|
:align: right
|
||||||
:alt: Add permissions for user
|
:alt: Add permissions for user
|
||||||
@ -236,7 +250,8 @@ You can list the ACLs of each user/token using the following command:
|
|||||||
│ john@pbs │ /datastore/store1 │ 1 │ DatastoreAdmin │
|
│ john@pbs │ /datastore/store1 │ 1 │ DatastoreAdmin │
|
||||||
└──────────┴───────────────────┴───────────┴────────────────┘
|
└──────────┴───────────────────┴───────────┴────────────────┘
|
||||||
|
|
||||||
A single user/token can be assigned multiple permission sets for different datastores.
|
A single user/token can be assigned multiple permission sets for different
|
||||||
|
datastores.
|
||||||
|
|
||||||
.. Note::
|
.. Note::
|
||||||
Naming convention is important here. For datastores on the host,
|
Naming convention is important here. For datastores on the host,
|
||||||
@ -247,11 +262,11 @@ A single user/token can be assigned multiple permission sets for different datas
|
|||||||
remote (see `Remote` below) and ``{storename}`` is the name of the datastore on
|
remote (see `Remote` below) and ``{storename}`` is the name of the datastore on
|
||||||
the remote.
|
the remote.
|
||||||
|
|
||||||
API Token permissions
|
API Token Permissions
|
||||||
~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
API token permissions are calculated based on ACLs containing their ID
|
API token permissions are calculated based on ACLs containing their ID,
|
||||||
independent of those of their corresponding user. The resulting permission set
|
independently of those of their corresponding user. The resulting permission set
|
||||||
on a given path is then intersected with that of the corresponding user.
|
on a given path is then intersected with that of the corresponding user.
|
||||||
|
|
||||||
In practice this means:
|
In practice this means:
|
||||||
@ -259,17 +274,17 @@ In practice this means:
|
|||||||
#. API tokens require their own ACL entries
|
#. API tokens require their own ACL entries
|
||||||
#. API tokens can never do more than their corresponding user
|
#. API tokens can never do more than their corresponding user
|
||||||
|
|
||||||
Effective permissions
|
Effective Permissions
|
||||||
~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
To calculate and display the effective permission set of a user or API token
|
To calculate and display the effective permission set of a user or API token,
|
||||||
you can use the ``proxmox-backup-manager user permission`` command:
|
you can use the ``proxmox-backup-manager user permission`` command:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
# proxmox-backup-manager user permissions john@pbs --path /datastore/store1
|
# proxmox-backup-manager user permissions john@pbs --path /datastore/store1
|
||||||
Privileges with (*) have the propagate flag set
|
Privileges with (*) have the propagate flag set
|
||||||
|
|
||||||
Path: /datastore/store1
|
Path: /datastore/store1
|
||||||
- Datastore.Audit (*)
|
- Datastore.Audit (*)
|
||||||
- Datastore.Backup (*)
|
- Datastore.Backup (*)
|
||||||
@ -277,17 +292,17 @@ you can use the ``proxmox-backup-manager user permission`` command:
|
|||||||
- Datastore.Prune (*)
|
- Datastore.Prune (*)
|
||||||
- Datastore.Read (*)
|
- Datastore.Read (*)
|
||||||
- Datastore.Verify (*)
|
- Datastore.Verify (*)
|
||||||
|
|
||||||
# proxmox-backup-manager acl update /datastore/store1 DatastoreBackup --auth-id 'john@pbs!client1'
|
# proxmox-backup-manager acl update /datastore/store1 DatastoreBackup --auth-id 'john@pbs!client1'
|
||||||
# proxmox-backup-manager user permissions 'john@pbs!client1' --path /datastore/store1
|
# proxmox-backup-manager user permissions 'john@pbs!client1' --path /datastore/store1
|
||||||
Privileges with (*) have the propagate flag set
|
Privileges with (*) have the propagate flag set
|
||||||
|
|
||||||
Path: /datastore/store1
|
Path: /datastore/store1
|
||||||
- Datastore.Backup (*)
|
- Datastore.Backup (*)
|
||||||
|
|
||||||
.. _user_tfa:
|
.. _user_tfa:
|
||||||
|
|
||||||
Two-factor authentication
|
Two-Factor Authentication
|
||||||
-------------------------
|
-------------------------
|
||||||
|
|
||||||
Introduction
|
Introduction
|
||||||
@ -296,7 +311,7 @@ Introduction
|
|||||||
With simple authentication, only a password (single factor) is required to
|
With simple authentication, only a password (single factor) is required to
|
||||||
successfully claim an identity (authenticate), for example, to be able to log in
|
successfully claim an identity (authenticate), for example, to be able to log in
|
||||||
as `root@pam` on a specific instance of Proxmox Backup Server. In this case, if
|
as `root@pam` on a specific instance of Proxmox Backup Server. In this case, if
|
||||||
the password gets stolen or leaked, anybody can use it to log in - even if they
|
the password gets leaked or stolen, anybody can use it to log in - even if they
|
||||||
should not be allowed to do so.
|
should not be allowed to do so.
|
||||||
|
|
||||||
With two-factor authentication (TFA), a user is asked for an additional factor
|
With two-factor authentication (TFA), a user is asked for an additional factor
|
||||||
@ -359,14 +374,18 @@ WebAuthn
|
|||||||
|
|
||||||
For WebAuthn to work, you need to have two things:
|
For WebAuthn to work, you need to have two things:
|
||||||
|
|
||||||
* a trusted HTTPS certificate (for example, by using `Let's Encrypt
|
* A trusted HTTPS certificate (for example, by using `Let's Encrypt
|
||||||
<https://pbs.proxmox.com/wiki/index.php/HTTPS_Certificate_Configuration>`_)
|
<https://pbs.proxmox.com/wiki/index.php/HTTPS_Certificate_Configuration>`_).
|
||||||
|
While it probably works with an untrusted certificate, some browsers may warn
|
||||||
|
or refuse WebAuthn operations if it is not trusted.
|
||||||
|
|
||||||
* setup the WebAuthn configuration (see *Configuration -> Authentication* in the
|
* Setup the WebAuthn configuration (see **Configuration -> Authentication** in
|
||||||
Proxmox Backup Server web-interface). This can be auto-filled in most setups.
|
the Proxmox Backup Server web interface). This can be auto-filled in most
|
||||||
|
setups.
|
||||||
|
|
||||||
Once you have fulfilled both of these requirements, you can add a WebAuthn
|
Once you have fulfilled both of these requirements, you can add a WebAuthn
|
||||||
configuration in the *Access Control* panel.
|
configuration in the **Two Factor Authentication** tab of the **Access Control**
|
||||||
|
panel.
|
||||||
|
|
||||||
.. _user_tfa_setup_recovery_keys:
|
.. _user_tfa_setup_recovery_keys:
|
||||||
|
|
||||||
@ -378,7 +397,8 @@ Recovery Keys
|
|||||||
:alt: Add a new user
|
:alt: Add a new user
|
||||||
|
|
||||||
Recovery key codes do not need any preparation; you can simply create a set of
|
Recovery key codes do not need any preparation; you can simply create a set of
|
||||||
recovery keys in the *Access Control* panel.
|
recovery keys in the **Two Factor Authentication** tab of the **Access Control**
|
||||||
|
panel.
|
||||||
|
|
||||||
.. note:: There can only be one set of single-use recovery keys per user at any
|
.. note:: There can only be one set of single-use recovery keys per user at any
|
||||||
time.
|
time.
|
||||||
|
@ -1 +1 @@
|
|||||||
deb https://enterprise.proxmox.com/debian/pbs buster pbs-enterprise
|
deb https://enterprise.proxmox.com/debian/pbs bullseye pbs-enterprise
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
use anyhow::{Error};
|
use anyhow::Error;
|
||||||
|
|
||||||
// chacha20-poly1305
|
// chacha20-poly1305
|
||||||
|
|
||||||
|
@ -1,6 +1,7 @@
|
|||||||
use anyhow::{Error};
|
use anyhow::{Error};
|
||||||
|
|
||||||
use proxmox::api::{*, cli::*};
|
use proxmox_schema::*;
|
||||||
|
use proxmox_router::cli::*;
|
||||||
|
|
||||||
#[api(
|
#[api(
|
||||||
input: {
|
input: {
|
||||||
|
@ -1,9 +1,9 @@
|
|||||||
use std::io::Write;
|
use std::io::Write;
|
||||||
|
|
||||||
use anyhow::{Error};
|
use anyhow::Error;
|
||||||
|
|
||||||
use proxmox_backup::api2::types::Authid;
|
use pbs_api_types::Authid;
|
||||||
use proxmox_backup::client::{HttpClient, HttpClientOptions, BackupReader};
|
use pbs_client::{HttpClient, HttpClientOptions, BackupReader};
|
||||||
|
|
||||||
pub struct DummyWriter {
|
pub struct DummyWriter {
|
||||||
bytes: usize,
|
bytes: usize,
|
||||||
@ -34,7 +34,7 @@ async fn run() -> Result<(), Error> {
|
|||||||
|
|
||||||
let client = HttpClient::new(host, 8007, auth_id, options)?;
|
let client = HttpClient::new(host, 8007, auth_id, options)?;
|
||||||
|
|
||||||
let backup_time = proxmox::tools::time::parse_rfc3339("2019-06-28T10:49:48Z")?;
|
let backup_time = proxmox_time::parse_rfc3339("2019-06-28T10:49:48Z")?;
|
||||||
|
|
||||||
let client = BackupReader::start(client, None, "store2", "host", "elsa", backup_time, true)
|
let client = BackupReader::start(client, None, "store2", "host", "elsa", backup_time, true)
|
||||||
.await?;
|
.await?;
|
||||||
@ -59,7 +59,7 @@ async fn run() -> Result<(), Error> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
if let Err(err) = proxmox_backup::tools::runtime::main(run()) {
|
if let Err(err) = proxmox_async::runtime::main(run()) {
|
||||||
eprintln!("ERROR: {}", err);
|
eprintln!("ERROR: {}", err);
|
||||||
}
|
}
|
||||||
println!("DONE");
|
println!("DONE");
|
||||||
|
@ -1,9 +1,9 @@
|
|||||||
use anyhow::{bail, Error};
|
|
||||||
|
|
||||||
use std::thread;
|
use std::thread;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
use std::io::Write;
|
use std::io::Write;
|
||||||
|
|
||||||
|
use anyhow::{bail, Error};
|
||||||
|
|
||||||
// tar handle files that shrink during backup, by simply padding with zeros.
|
// tar handle files that shrink during backup, by simply padding with zeros.
|
||||||
//
|
//
|
||||||
// this binary run multiple thread which writes some large files, then truncates
|
// this binary run multiple thread which writes some large files, then truncates
|
||||||
|
@ -69,7 +69,7 @@ fn send_request(
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn main() -> Result<(), Error> {
|
fn main() -> Result<(), Error> {
|
||||||
proxmox_backup::tools::runtime::main(run())
|
proxmox_async::runtime::main(run())
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn run() -> Result<(), Error> {
|
async fn run() -> Result<(), Error> {
|
||||||
|
@ -69,7 +69,7 @@ fn send_request(
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn main() -> Result<(), Error> {
|
fn main() -> Result<(), Error> {
|
||||||
proxmox_backup::tools::runtime::main(run())
|
proxmox_async::runtime::main(run())
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn run() -> Result<(), Error> {
|
async fn run() -> Result<(), Error> {
|
||||||
|
@ -6,10 +6,10 @@ use hyper::{Body, Request, Response};
|
|||||||
use openssl::ssl::{SslAcceptor, SslFiletype, SslMethod};
|
use openssl::ssl::{SslAcceptor, SslFiletype, SslMethod};
|
||||||
use tokio::net::{TcpListener, TcpStream};
|
use tokio::net::{TcpListener, TcpStream};
|
||||||
|
|
||||||
use proxmox_backup::configdir;
|
use pbs_buildcfg::configdir;
|
||||||
|
|
||||||
fn main() -> Result<(), Error> {
|
fn main() -> Result<(), Error> {
|
||||||
proxmox_backup::tools::runtime::main(run())
|
proxmox_async::runtime::main(run())
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn run() -> Result<(), Error> {
|
async fn run() -> Result<(), Error> {
|
||||||
|
@ -5,7 +5,7 @@ use hyper::{Body, Request, Response};
|
|||||||
use tokio::net::{TcpListener, TcpStream};
|
use tokio::net::{TcpListener, TcpStream};
|
||||||
|
|
||||||
fn main() -> Result<(), Error> {
|
fn main() -> Result<(), Error> {
|
||||||
proxmox_backup::tools::runtime::main(run())
|
proxmox_async::runtime::main(run())
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn run() -> Result<(), Error> {
|
async fn run() -> Result<(), Error> {
|
||||||
|
@ -5,7 +5,7 @@ extern crate proxmox_backup;
|
|||||||
use anyhow::{Error};
|
use anyhow::{Error};
|
||||||
use std::io::{Read, Write};
|
use std::io::{Read, Write};
|
||||||
|
|
||||||
use proxmox_backup::backup::*;
|
use pbs_datastore::Chunker;
|
||||||
|
|
||||||
struct ChunkWriter {
|
struct ChunkWriter {
|
||||||
chunker: Chunker,
|
chunker: Chunker,
|
||||||
|
@ -1,7 +1,6 @@
|
|||||||
extern crate proxmox_backup;
|
extern crate proxmox_backup;
|
||||||
|
|
||||||
//use proxmox_backup::backup::chunker::*;
|
use pbs_datastore::Chunker;
|
||||||
use proxmox_backup::backup::*;
|
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
|
|
||||||
|
@ -3,7 +3,7 @@ use futures::*;
|
|||||||
|
|
||||||
extern crate proxmox_backup;
|
extern crate proxmox_backup;
|
||||||
|
|
||||||
use proxmox_backup::backup::*;
|
use pbs_client::ChunkStream;
|
||||||
|
|
||||||
// Test Chunker with real data read from a file.
|
// Test Chunker with real data read from a file.
|
||||||
//
|
//
|
||||||
@ -13,7 +13,7 @@ use proxmox_backup::backup::*;
|
|||||||
// Note: I can currently get about 830MB/s
|
// Note: I can currently get about 830MB/s
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
if let Err(err) = proxmox_backup::tools::runtime::main(run()) {
|
if let Err(err) = proxmox_async::runtime::main(run()) {
|
||||||
panic!("ERROR: {}", err);
|
panic!("ERROR: {}", err);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
use anyhow::{Error};
|
use anyhow::{Error};
|
||||||
|
|
||||||
use proxmox_backup::api2::types::Authid;
|
use pbs_client::{HttpClient, HttpClientOptions, BackupWriter};
|
||||||
use proxmox_backup::client::*;
|
use pbs_api_types::Authid;
|
||||||
|
|
||||||
async fn upload_speed() -> Result<f64, Error> {
|
async fn upload_speed() -> Result<f64, Error> {
|
||||||
|
|
||||||
@ -16,7 +16,7 @@ async fn upload_speed() -> Result<f64, Error> {
|
|||||||
|
|
||||||
let client = HttpClient::new(host, 8007, auth_id, options)?;
|
let client = HttpClient::new(host, 8007, auth_id, options)?;
|
||||||
|
|
||||||
let backup_time = proxmox::tools::time::epoch_i64();
|
let backup_time = proxmox_time::epoch_i64();
|
||||||
|
|
||||||
let client = BackupWriter::start(client, None, datastore, "host", "speedtest", backup_time, false, true).await?;
|
let client = BackupWriter::start(client, None, datastore, "host", "speedtest", backup_time, false, true).await?;
|
||||||
|
|
||||||
@ -27,7 +27,7 @@ async fn upload_speed() -> Result<f64, Error> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
match proxmox_backup::tools::runtime::main(upload_speed()) {
|
match proxmox_async::runtime::main(upload_speed()) {
|
||||||
Ok(mbs) => {
|
Ok(mbs) => {
|
||||||
println!("average upload speed: {} MB/s", mbs);
|
println!("average upload speed: {} MB/s", mbs);
|
||||||
}
|
}
|
||||||
|
22
pbs-api-types/Cargo.toml
Normal file
22
pbs-api-types/Cargo.toml
Normal file
@ -0,0 +1,22 @@
|
|||||||
|
[package]
|
||||||
|
name = "pbs-api-types"
|
||||||
|
version = "0.1.0"
|
||||||
|
authors = ["Proxmox Support Team <support@proxmox.com>"]
|
||||||
|
edition = "2018"
|
||||||
|
description = "general API type helpers for PBS"
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
anyhow = "1.0"
|
||||||
|
hex = "0.4.3"
|
||||||
|
lazy_static = "1.4"
|
||||||
|
libc = "0.2"
|
||||||
|
nix = "0.19.1"
|
||||||
|
openssl = "0.10"
|
||||||
|
regex = "1.2"
|
||||||
|
serde = { version = "1.0", features = ["derive"] }
|
||||||
|
|
||||||
|
proxmox = "0.15.3"
|
||||||
|
proxmox-lang = "1.0.0"
|
||||||
|
proxmox-schema = { version = "1.0.1", features = [ "api-macro" ] }
|
||||||
|
proxmox-time = "1.1"
|
||||||
|
proxmox-uuid = { version = "1.0.0", features = [ "serde" ] }
|
280
pbs-api-types/src/acl.rs
Normal file
280
pbs-api-types/src/acl.rs
Normal file
@ -0,0 +1,280 @@
|
|||||||
|
use std::str::FromStr;
|
||||||
|
|
||||||
|
use serde::de::{value, IntoDeserializer};
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_lang::constnamedbitmap;
|
||||||
|
use proxmox_schema::{
|
||||||
|
api, const_regex, ApiStringFormat, BooleanSchema, EnumEntry, Schema, StringSchema,
|
||||||
|
};
|
||||||
|
|
||||||
|
const_regex! {
|
||||||
|
pub ACL_PATH_REGEX = concat!(r"^(?:/|", r"(?:/", PROXMOX_SAFE_ID_REGEX_STR!(), ")+", r")$");
|
||||||
|
}
|
||||||
|
|
||||||
|
// define Privilege bitfield
|
||||||
|
|
||||||
|
constnamedbitmap! {
|
||||||
|
/// Contains a list of privilege name to privilege value mappings.
|
||||||
|
///
|
||||||
|
/// The names are used when displaying/persisting privileges anywhere, the values are used to
|
||||||
|
/// allow easy matching of privileges as bitflags.
|
||||||
|
PRIVILEGES: u64 => {
|
||||||
|
/// Sys.Audit allows knowing about the system and its status
|
||||||
|
PRIV_SYS_AUDIT("Sys.Audit");
|
||||||
|
/// Sys.Modify allows modifying system-level configuration
|
||||||
|
PRIV_SYS_MODIFY("Sys.Modify");
|
||||||
|
/// Sys.Modify allows to poweroff/reboot/.. the system
|
||||||
|
PRIV_SYS_POWER_MANAGEMENT("Sys.PowerManagement");
|
||||||
|
|
||||||
|
/// Datastore.Audit allows knowing about a datastore,
|
||||||
|
/// including reading the configuration entry and listing its contents
|
||||||
|
PRIV_DATASTORE_AUDIT("Datastore.Audit");
|
||||||
|
/// Datastore.Allocate allows creating or deleting datastores
|
||||||
|
PRIV_DATASTORE_ALLOCATE("Datastore.Allocate");
|
||||||
|
/// Datastore.Modify allows modifying a datastore and its contents
|
||||||
|
PRIV_DATASTORE_MODIFY("Datastore.Modify");
|
||||||
|
/// Datastore.Read allows reading arbitrary backup contents
|
||||||
|
PRIV_DATASTORE_READ("Datastore.Read");
|
||||||
|
/// Allows verifying a datastore
|
||||||
|
PRIV_DATASTORE_VERIFY("Datastore.Verify");
|
||||||
|
|
||||||
|
/// Datastore.Backup allows Datastore.Read|Verify and creating new snapshots,
|
||||||
|
/// but also requires backup ownership
|
||||||
|
PRIV_DATASTORE_BACKUP("Datastore.Backup");
|
||||||
|
/// Datastore.Prune allows deleting snapshots,
|
||||||
|
/// but also requires backup ownership
|
||||||
|
PRIV_DATASTORE_PRUNE("Datastore.Prune");
|
||||||
|
|
||||||
|
/// Permissions.Modify allows modifying ACLs
|
||||||
|
PRIV_PERMISSIONS_MODIFY("Permissions.Modify");
|
||||||
|
|
||||||
|
/// Remote.Audit allows reading remote.cfg and sync.cfg entries
|
||||||
|
PRIV_REMOTE_AUDIT("Remote.Audit");
|
||||||
|
/// Remote.Modify allows modifying remote.cfg
|
||||||
|
PRIV_REMOTE_MODIFY("Remote.Modify");
|
||||||
|
/// Remote.Read allows reading data from a configured `Remote`
|
||||||
|
PRIV_REMOTE_READ("Remote.Read");
|
||||||
|
|
||||||
|
/// Sys.Console allows access to the system's console
|
||||||
|
PRIV_SYS_CONSOLE("Sys.Console");
|
||||||
|
|
||||||
|
/// Tape.Audit allows reading tape backup configuration and status
|
||||||
|
PRIV_TAPE_AUDIT("Tape.Audit");
|
||||||
|
/// Tape.Modify allows modifying tape backup configuration
|
||||||
|
PRIV_TAPE_MODIFY("Tape.Modify");
|
||||||
|
/// Tape.Write allows writing tape media
|
||||||
|
PRIV_TAPE_WRITE("Tape.Write");
|
||||||
|
/// Tape.Read allows reading tape backup configuration and media contents
|
||||||
|
PRIV_TAPE_READ("Tape.Read");
|
||||||
|
|
||||||
|
/// Realm.Allocate allows viewing, creating, modifying and deleting realms
|
||||||
|
PRIV_REALM_ALLOCATE("Realm.Allocate");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Admin always has all privileges. It can do everything except a few actions
|
||||||
|
/// which are limited to the 'root@pam` superuser
|
||||||
|
pub const ROLE_ADMIN: u64 = u64::MAX;
|
||||||
|
|
||||||
|
/// NoAccess can be used to remove privileges from specific (sub-)paths
|
||||||
|
pub const ROLE_NO_ACCESS: u64 = 0;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Audit can view configuration and status information, but not modify it.
|
||||||
|
pub const ROLE_AUDIT: u64 = 0
|
||||||
|
| PRIV_SYS_AUDIT
|
||||||
|
| PRIV_DATASTORE_AUDIT;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Datastore.Admin can do anything on the datastore.
|
||||||
|
pub const ROLE_DATASTORE_ADMIN: u64 = 0
|
||||||
|
| PRIV_DATASTORE_AUDIT
|
||||||
|
| PRIV_DATASTORE_MODIFY
|
||||||
|
| PRIV_DATASTORE_READ
|
||||||
|
| PRIV_DATASTORE_VERIFY
|
||||||
|
| PRIV_DATASTORE_BACKUP
|
||||||
|
| PRIV_DATASTORE_PRUNE;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Datastore.Reader can read/verify datastore content and do restore
|
||||||
|
pub const ROLE_DATASTORE_READER: u64 = 0
|
||||||
|
| PRIV_DATASTORE_AUDIT
|
||||||
|
| PRIV_DATASTORE_VERIFY
|
||||||
|
| PRIV_DATASTORE_READ;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Datastore.Backup can do backup and restore, but no prune.
|
||||||
|
pub const ROLE_DATASTORE_BACKUP: u64 = 0
|
||||||
|
| PRIV_DATASTORE_BACKUP;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Datastore.PowerUser can do backup, restore, and prune.
|
||||||
|
pub const ROLE_DATASTORE_POWERUSER: u64 = 0
|
||||||
|
| PRIV_DATASTORE_PRUNE
|
||||||
|
| PRIV_DATASTORE_BACKUP;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Datastore.Audit can audit the datastore.
|
||||||
|
pub const ROLE_DATASTORE_AUDIT: u64 = 0
|
||||||
|
| PRIV_DATASTORE_AUDIT;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Remote.Audit can audit the remote
|
||||||
|
pub const ROLE_REMOTE_AUDIT: u64 = 0
|
||||||
|
| PRIV_REMOTE_AUDIT;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Remote.Admin can do anything on the remote.
|
||||||
|
pub const ROLE_REMOTE_ADMIN: u64 = 0
|
||||||
|
| PRIV_REMOTE_AUDIT
|
||||||
|
| PRIV_REMOTE_MODIFY
|
||||||
|
| PRIV_REMOTE_READ;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Remote.SyncOperator can do read and prune on the remote.
|
||||||
|
pub const ROLE_REMOTE_SYNC_OPERATOR: u64 = 0
|
||||||
|
| PRIV_REMOTE_AUDIT
|
||||||
|
| PRIV_REMOTE_READ;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Tape.Audit can audit the tape backup configuration and media content
|
||||||
|
pub const ROLE_TAPE_AUDIT: u64 = 0
|
||||||
|
| PRIV_TAPE_AUDIT;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Tape.Admin can do anything on the tape backup
|
||||||
|
pub const ROLE_TAPE_ADMIN: u64 = 0
|
||||||
|
| PRIV_TAPE_AUDIT
|
||||||
|
| PRIV_TAPE_MODIFY
|
||||||
|
| PRIV_TAPE_READ
|
||||||
|
| PRIV_TAPE_WRITE;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Tape.Operator can do tape backup and restore (but no configuration changes)
|
||||||
|
pub const ROLE_TAPE_OPERATOR: u64 = 0
|
||||||
|
| PRIV_TAPE_AUDIT
|
||||||
|
| PRIV_TAPE_READ
|
||||||
|
| PRIV_TAPE_WRITE;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[allow(clippy::identity_op)]
|
||||||
|
/// Tape.Reader can do read and inspect tape content
|
||||||
|
pub const ROLE_TAPE_READER: u64 = 0
|
||||||
|
| PRIV_TAPE_AUDIT
|
||||||
|
| PRIV_TAPE_READ;
|
||||||
|
|
||||||
|
/// NoAccess can be used to remove privileges from specific (sub-)paths
|
||||||
|
pub const ROLE_NAME_NO_ACCESS: &str = "NoAccess";
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
type_text: "<role>",
|
||||||
|
)]
|
||||||
|
#[repr(u64)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
/// Enum representing roles via their [PRIVILEGES] combination.
|
||||||
|
///
|
||||||
|
/// Since privileges are implemented as bitflags, each unique combination of privileges maps to a
|
||||||
|
/// single, unique `u64` value that is used in this enum definition.
|
||||||
|
pub enum Role {
|
||||||
|
/// Administrator
|
||||||
|
Admin = ROLE_ADMIN,
|
||||||
|
/// Auditor
|
||||||
|
Audit = ROLE_AUDIT,
|
||||||
|
/// Disable Access
|
||||||
|
NoAccess = ROLE_NO_ACCESS,
|
||||||
|
/// Datastore Administrator
|
||||||
|
DatastoreAdmin = ROLE_DATASTORE_ADMIN,
|
||||||
|
/// Datastore Reader (inspect datastore content and do restores)
|
||||||
|
DatastoreReader = ROLE_DATASTORE_READER,
|
||||||
|
/// Datastore Backup (backup and restore owned backups)
|
||||||
|
DatastoreBackup = ROLE_DATASTORE_BACKUP,
|
||||||
|
/// Datastore PowerUser (backup, restore and prune owned backup)
|
||||||
|
DatastorePowerUser = ROLE_DATASTORE_POWERUSER,
|
||||||
|
/// Datastore Auditor
|
||||||
|
DatastoreAudit = ROLE_DATASTORE_AUDIT,
|
||||||
|
/// Remote Auditor
|
||||||
|
RemoteAudit = ROLE_REMOTE_AUDIT,
|
||||||
|
/// Remote Administrator
|
||||||
|
RemoteAdmin = ROLE_REMOTE_ADMIN,
|
||||||
|
/// Syncronisation Opertator
|
||||||
|
RemoteSyncOperator = ROLE_REMOTE_SYNC_OPERATOR,
|
||||||
|
/// Tape Auditor
|
||||||
|
TapeAudit = ROLE_TAPE_AUDIT,
|
||||||
|
/// Tape Administrator
|
||||||
|
TapeAdmin = ROLE_TAPE_ADMIN,
|
||||||
|
/// Tape Operator
|
||||||
|
TapeOperator = ROLE_TAPE_OPERATOR,
|
||||||
|
/// Tape Reader
|
||||||
|
TapeReader = ROLE_TAPE_READER,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl FromStr for Role {
|
||||||
|
type Err = value::Error;
|
||||||
|
|
||||||
|
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||||
|
Self::deserialize(s.into_deserializer())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const ACL_PATH_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&ACL_PATH_REGEX);
|
||||||
|
|
||||||
|
pub const ACL_PATH_SCHEMA: Schema = StringSchema::new("Access control path.")
|
||||||
|
.format(&ACL_PATH_FORMAT)
|
||||||
|
.min_length(1)
|
||||||
|
.max_length(128)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const ACL_PROPAGATE_SCHEMA: Schema =
|
||||||
|
BooleanSchema::new("Allow to propagate (inherit) permissions.")
|
||||||
|
.default(true)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const ACL_UGID_TYPE_SCHEMA: Schema = StringSchema::new("Type of 'ugid' property.")
|
||||||
|
.format(&ApiStringFormat::Enum(&[
|
||||||
|
EnumEntry::new("user", "User"),
|
||||||
|
EnumEntry::new("group", "Group"),
|
||||||
|
]))
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
propagate: {
|
||||||
|
schema: ACL_PROPAGATE_SCHEMA,
|
||||||
|
},
|
||||||
|
path: {
|
||||||
|
schema: ACL_PATH_SCHEMA,
|
||||||
|
},
|
||||||
|
ugid_type: {
|
||||||
|
schema: ACL_UGID_TYPE_SCHEMA,
|
||||||
|
},
|
||||||
|
ugid: {
|
||||||
|
type: String,
|
||||||
|
description: "User or Group ID.",
|
||||||
|
},
|
||||||
|
roleid: {
|
||||||
|
type: Role,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
/// ACL list entry.
|
||||||
|
pub struct AclListItem {
|
||||||
|
pub path: String,
|
||||||
|
pub ugid: String,
|
||||||
|
pub ugid_type: String,
|
||||||
|
pub propagate: bool,
|
||||||
|
pub roleid: String,
|
||||||
|
}
|
98
pbs-api-types/src/crypto.rs
Normal file
98
pbs-api-types/src/crypto.rs
Normal file
@ -0,0 +1,98 @@
|
|||||||
|
use std::fmt::{self, Display};
|
||||||
|
|
||||||
|
use anyhow::Error;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::api;
|
||||||
|
|
||||||
|
#[api(default: "encrypt")]
|
||||||
|
#[derive(Copy, Clone, Debug, Eq, PartialEq, Deserialize, Serialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Defines whether data is encrypted (using an AEAD cipher), only signed, or neither.
|
||||||
|
pub enum CryptMode {
|
||||||
|
/// Don't encrypt.
|
||||||
|
None,
|
||||||
|
/// Encrypt.
|
||||||
|
Encrypt,
|
||||||
|
/// Only sign.
|
||||||
|
SignOnly,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Eq, PartialEq, Hash, Clone, Deserialize, Serialize)]
|
||||||
|
#[serde(transparent)]
|
||||||
|
/// 32-byte fingerprint, usually calculated with SHA256.
|
||||||
|
pub struct Fingerprint {
|
||||||
|
#[serde(with = "bytes_as_fingerprint")]
|
||||||
|
bytes: [u8; 32],
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Fingerprint {
|
||||||
|
pub fn new(bytes: [u8; 32]) -> Self {
|
||||||
|
Self { bytes }
|
||||||
|
}
|
||||||
|
pub fn bytes(&self) -> &[u8; 32] {
|
||||||
|
&self.bytes
|
||||||
|
}
|
||||||
|
pub fn signature(&self) -> String {
|
||||||
|
as_fingerprint(&self.bytes)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Display as short key ID
|
||||||
|
impl Display for Fingerprint {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
|
write!(f, "{}", as_fingerprint(&self.bytes[0..8]))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::str::FromStr for Fingerprint {
|
||||||
|
type Err = Error;
|
||||||
|
|
||||||
|
fn from_str(s: &str) -> Result<Self, Error> {
|
||||||
|
let mut tmp = s.to_string();
|
||||||
|
tmp.retain(|c| c != ':');
|
||||||
|
let bytes = proxmox::tools::hex_to_digest(&tmp)?;
|
||||||
|
Ok(Fingerprint::new(bytes))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn as_fingerprint(bytes: &[u8]) -> String {
|
||||||
|
hex::encode(bytes)
|
||||||
|
.as_bytes()
|
||||||
|
.chunks(2)
|
||||||
|
.map(|v| unsafe { std::str::from_utf8_unchecked(v) }) // it's a hex string
|
||||||
|
.collect::<Vec<&str>>().join(":")
|
||||||
|
}
|
||||||
|
|
||||||
|
pub mod bytes_as_fingerprint {
|
||||||
|
use std::mem::MaybeUninit;
|
||||||
|
|
||||||
|
use serde::{Deserialize, Serializer, Deserializer};
|
||||||
|
|
||||||
|
pub fn serialize<S>(
|
||||||
|
bytes: &[u8; 32],
|
||||||
|
serializer: S,
|
||||||
|
) -> Result<S::Ok, S::Error>
|
||||||
|
where
|
||||||
|
S: Serializer,
|
||||||
|
{
|
||||||
|
let s = super::as_fingerprint(bytes);
|
||||||
|
serializer.serialize_str(&s)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn deserialize<'de, D>(
|
||||||
|
deserializer: D,
|
||||||
|
) -> Result<[u8; 32], D::Error>
|
||||||
|
where
|
||||||
|
D: Deserializer<'de>,
|
||||||
|
{
|
||||||
|
// TODO: more efficiently implement with a Visitor implementing visit_str using split() and
|
||||||
|
// hex::decode by-byte
|
||||||
|
let mut s = String::deserialize(deserializer)?;
|
||||||
|
s.retain(|c| c != ':');
|
||||||
|
let mut out = MaybeUninit::<[u8; 32]>::uninit();
|
||||||
|
hex::decode_to_slice(s.as_bytes(), unsafe { &mut (*out.as_mut_ptr())[..] })
|
||||||
|
.map_err(serde::de::Error::custom)?;
|
||||||
|
Ok(unsafe { out.assume_init() })
|
||||||
|
}
|
||||||
|
}
|
627
pbs-api-types/src/datastore.rs
Normal file
627
pbs-api-types/src/datastore.rs
Normal file
@ -0,0 +1,627 @@
|
|||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::{
|
||||||
|
api, const_regex, ApiStringFormat, ApiType, ArraySchema, EnumEntry, IntegerSchema, ReturnType,
|
||||||
|
Schema, StringSchema, Updater,
|
||||||
|
};
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
PROXMOX_SAFE_ID_FORMAT, SHA256_HEX_REGEX, SINGLE_LINE_COMMENT_SCHEMA, CryptMode, UPID,
|
||||||
|
Fingerprint, Userid, Authid,
|
||||||
|
GC_SCHEDULE_SCHEMA, DATASTORE_NOTIFY_STRING_SCHEMA, PRUNE_SCHEDULE_SCHEMA,
|
||||||
|
|
||||||
|
};
|
||||||
|
|
||||||
|
const_regex!{
|
||||||
|
pub BACKUP_TYPE_REGEX = concat!(r"^(", BACKUP_TYPE_RE!(), r")$");
|
||||||
|
|
||||||
|
pub BACKUP_ID_REGEX = concat!(r"^", BACKUP_ID_RE!(), r"$");
|
||||||
|
|
||||||
|
pub BACKUP_DATE_REGEX = concat!(r"^", BACKUP_TIME_RE!() ,r"$");
|
||||||
|
|
||||||
|
pub GROUP_PATH_REGEX = concat!(r"^(", BACKUP_TYPE_RE!(), ")/(", BACKUP_ID_RE!(), r")$");
|
||||||
|
|
||||||
|
pub BACKUP_FILE_REGEX = r"^.*\.([fd]idx|blob)$";
|
||||||
|
|
||||||
|
pub SNAPSHOT_PATH_REGEX = concat!(r"^", SNAPSHOT_PATH_REGEX_STR!(), r"$");
|
||||||
|
|
||||||
|
pub DATASTORE_MAP_REGEX = concat!(r"(:?", PROXMOX_SAFE_ID_REGEX_STR!(), r"=)?", PROXMOX_SAFE_ID_REGEX_STR!());
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const CHUNK_DIGEST_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&SHA256_HEX_REGEX);
|
||||||
|
|
||||||
|
pub const DIR_NAME_SCHEMA: Schema = StringSchema::new("Directory name")
|
||||||
|
.min_length(1)
|
||||||
|
.max_length(4096)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const BACKUP_ARCHIVE_NAME_SCHEMA: Schema = StringSchema::new("Backup archive name.")
|
||||||
|
.format(&PROXMOX_SAFE_ID_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const BACKUP_ID_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&BACKUP_ID_REGEX);
|
||||||
|
pub const BACKUP_GROUP_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&GROUP_PATH_REGEX);
|
||||||
|
|
||||||
|
pub const BACKUP_ID_SCHEMA: Schema = StringSchema::new("Backup ID.")
|
||||||
|
.format(&BACKUP_ID_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const BACKUP_TYPE_SCHEMA: Schema = StringSchema::new("Backup type.")
|
||||||
|
.format(&ApiStringFormat::Enum(&[
|
||||||
|
EnumEntry::new("vm", "Virtual Machine Backup"),
|
||||||
|
EnumEntry::new("ct", "Container Backup"),
|
||||||
|
EnumEntry::new("host", "Host Backup"),
|
||||||
|
]))
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const BACKUP_TIME_SCHEMA: Schema = IntegerSchema::new("Backup time (Unix epoch.)")
|
||||||
|
.minimum(1_547_797_308)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const BACKUP_GROUP_SCHEMA: Schema = StringSchema::new("Backup Group")
|
||||||
|
.format(&BACKUP_GROUP_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const DATASTORE_SCHEMA: Schema = StringSchema::new("Datastore name.")
|
||||||
|
.format(&PROXMOX_SAFE_ID_FORMAT)
|
||||||
|
.min_length(3)
|
||||||
|
.max_length(32)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const CHUNK_DIGEST_SCHEMA: Schema = StringSchema::new("Chunk digest (SHA256).")
|
||||||
|
.format(&CHUNK_DIGEST_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const DATASTORE_MAP_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&DATASTORE_MAP_REGEX);
|
||||||
|
|
||||||
|
pub const DATASTORE_MAP_SCHEMA: Schema = StringSchema::new("Datastore mapping.")
|
||||||
|
.format(&DATASTORE_MAP_FORMAT)
|
||||||
|
.min_length(3)
|
||||||
|
.max_length(65)
|
||||||
|
.type_text("(<source>=)?<target>")
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const DATASTORE_MAP_ARRAY_SCHEMA: Schema = ArraySchema::new(
|
||||||
|
"Datastore mapping list.", &DATASTORE_MAP_SCHEMA)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const DATASTORE_MAP_LIST_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"A list of Datastore mappings (or single datastore), comma separated. \
|
||||||
|
For example 'a=b,e' maps the source datastore 'a' to target 'b and \
|
||||||
|
all other sources to the default 'e'. If no default is given, only the \
|
||||||
|
specified sources are mapped.")
|
||||||
|
.format(&ApiStringFormat::PropertyString(&DATASTORE_MAP_ARRAY_SCHEMA))
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PRUNE_SCHEMA_KEEP_DAILY: Schema = IntegerSchema::new("Number of daily backups to keep.")
|
||||||
|
.minimum(1)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PRUNE_SCHEMA_KEEP_HOURLY: Schema =
|
||||||
|
IntegerSchema::new("Number of hourly backups to keep.")
|
||||||
|
.minimum(1)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PRUNE_SCHEMA_KEEP_LAST: Schema = IntegerSchema::new("Number of backups to keep.")
|
||||||
|
.minimum(1)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PRUNE_SCHEMA_KEEP_MONTHLY: Schema =
|
||||||
|
IntegerSchema::new("Number of monthly backups to keep.")
|
||||||
|
.minimum(1)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PRUNE_SCHEMA_KEEP_WEEKLY: Schema =
|
||||||
|
IntegerSchema::new("Number of weekly backups to keep.")
|
||||||
|
.minimum(1)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PRUNE_SCHEMA_KEEP_YEARLY: Schema =
|
||||||
|
IntegerSchema::new("Number of yearly backups to keep.")
|
||||||
|
.minimum(1)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"keep-last": {
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_LAST,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"keep-hourly": {
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_HOURLY,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"keep-daily": {
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_DAILY,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"keep-weekly": {
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_WEEKLY,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"keep-monthly": {
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_MONTHLY,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"keep-yearly": {
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_YEARLY,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize, Default)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Common pruning options
|
||||||
|
pub struct PruneOptions {
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_last: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_hourly: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_daily: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_weekly: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_monthly: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_yearly: Option<u64>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
name: {
|
||||||
|
schema: DATASTORE_SCHEMA,
|
||||||
|
},
|
||||||
|
path: {
|
||||||
|
schema: DIR_NAME_SCHEMA,
|
||||||
|
},
|
||||||
|
"notify-user": {
|
||||||
|
optional: true,
|
||||||
|
type: Userid,
|
||||||
|
},
|
||||||
|
"notify": {
|
||||||
|
optional: true,
|
||||||
|
schema: DATASTORE_NOTIFY_STRING_SCHEMA,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
"gc-schedule": {
|
||||||
|
optional: true,
|
||||||
|
schema: GC_SCHEDULE_SCHEMA,
|
||||||
|
},
|
||||||
|
"prune-schedule": {
|
||||||
|
optional: true,
|
||||||
|
schema: PRUNE_SCHEDULE_SCHEMA,
|
||||||
|
},
|
||||||
|
"keep-last": {
|
||||||
|
optional: true,
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_LAST,
|
||||||
|
},
|
||||||
|
"keep-hourly": {
|
||||||
|
optional: true,
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_HOURLY,
|
||||||
|
},
|
||||||
|
"keep-daily": {
|
||||||
|
optional: true,
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_DAILY,
|
||||||
|
},
|
||||||
|
"keep-weekly": {
|
||||||
|
optional: true,
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_WEEKLY,
|
||||||
|
},
|
||||||
|
"keep-monthly": {
|
||||||
|
optional: true,
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_MONTHLY,
|
||||||
|
},
|
||||||
|
"keep-yearly": {
|
||||||
|
optional: true,
|
||||||
|
schema: PRUNE_SCHEMA_KEEP_YEARLY,
|
||||||
|
},
|
||||||
|
"verify-new": {
|
||||||
|
description: "If enabled, all new backups will be verified right after completion.",
|
||||||
|
optional: true,
|
||||||
|
type: bool,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize,Updater)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Datastore configuration properties.
|
||||||
|
pub struct DataStoreConfig {
|
||||||
|
#[updater(skip)]
|
||||||
|
pub name: String,
|
||||||
|
#[updater(skip)]
|
||||||
|
pub path: String,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub gc_schedule: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub prune_schedule: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_last: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_hourly: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_daily: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_weekly: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_monthly: Option<u64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub keep_yearly: Option<u64>,
|
||||||
|
/// If enabled, all backups will be verified right after completion.
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub verify_new: Option<bool>,
|
||||||
|
/// Send job email notification to this user
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub notify_user: Option<Userid>,
|
||||||
|
/// Send notification only for job errors
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub notify: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
store: {
|
||||||
|
schema: DATASTORE_SCHEMA,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Basic information about a datastore.
|
||||||
|
pub struct DataStoreListItem {
|
||||||
|
pub store: String,
|
||||||
|
pub comment: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"filename": {
|
||||||
|
schema: BACKUP_ARCHIVE_NAME_SCHEMA,
|
||||||
|
},
|
||||||
|
"crypt-mode": {
|
||||||
|
type: CryptMode,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Basic information about archive files inside a backup snapshot.
|
||||||
|
pub struct BackupContent {
|
||||||
|
pub filename: String,
|
||||||
|
/// Info if file is encrypted, signed, or neither.
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub crypt_mode: Option<CryptMode>,
|
||||||
|
/// Archive size (from backup manifest).
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub size: Option<u64>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
/// Result of a verify operation.
|
||||||
|
pub enum VerifyState {
|
||||||
|
/// Verification was successful
|
||||||
|
Ok,
|
||||||
|
/// Verification reported one or more errors
|
||||||
|
Failed,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
upid: {
|
||||||
|
type: UPID,
|
||||||
|
},
|
||||||
|
state: {
|
||||||
|
type: VerifyState,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
/// Task properties.
|
||||||
|
pub struct SnapshotVerifyState {
|
||||||
|
/// UPID of the verify task
|
||||||
|
pub upid: UPID,
|
||||||
|
/// State of the verification. Enum.
|
||||||
|
pub state: VerifyState,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"backup-type": {
|
||||||
|
schema: BACKUP_TYPE_SCHEMA,
|
||||||
|
},
|
||||||
|
"backup-id": {
|
||||||
|
schema: BACKUP_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
"backup-time": {
|
||||||
|
schema: BACKUP_TIME_SCHEMA,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
verification: {
|
||||||
|
type: SnapshotVerifyState,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
fingerprint: {
|
||||||
|
type: String,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
files: {
|
||||||
|
items: {
|
||||||
|
schema: BACKUP_ARCHIVE_NAME_SCHEMA
|
||||||
|
},
|
||||||
|
},
|
||||||
|
owner: {
|
||||||
|
type: Authid,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Basic information about backup snapshot.
|
||||||
|
pub struct SnapshotListItem {
|
||||||
|
pub backup_type: String, // enum
|
||||||
|
pub backup_id: String,
|
||||||
|
pub backup_time: i64,
|
||||||
|
/// The first line from manifest "notes"
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
/// The result of the last run verify task
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub verification: Option<SnapshotVerifyState>,
|
||||||
|
/// Fingerprint of encryption key
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub fingerprint: Option<Fingerprint>,
|
||||||
|
/// List of contained archive files.
|
||||||
|
pub files: Vec<BackupContent>,
|
||||||
|
/// Overall snapshot size (sum of all archive sizes).
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub size: Option<u64>,
|
||||||
|
/// The owner of the snapshots group
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub owner: Option<Authid>,
|
||||||
|
/// Protection from prunes
|
||||||
|
#[serde(default)]
|
||||||
|
pub protected: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"backup-type": {
|
||||||
|
schema: BACKUP_TYPE_SCHEMA,
|
||||||
|
},
|
||||||
|
"backup-id": {
|
||||||
|
schema: BACKUP_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
"last-backup": {
|
||||||
|
schema: BACKUP_TIME_SCHEMA,
|
||||||
|
},
|
||||||
|
"backup-count": {
|
||||||
|
type: Integer,
|
||||||
|
},
|
||||||
|
files: {
|
||||||
|
items: {
|
||||||
|
schema: BACKUP_ARCHIVE_NAME_SCHEMA
|
||||||
|
},
|
||||||
|
},
|
||||||
|
owner: {
|
||||||
|
type: Authid,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Basic information about a backup group.
|
||||||
|
pub struct GroupListItem {
|
||||||
|
pub backup_type: String, // enum
|
||||||
|
pub backup_id: String,
|
||||||
|
pub last_backup: i64,
|
||||||
|
/// Number of contained snapshots
|
||||||
|
pub backup_count: u64,
|
||||||
|
/// List of contained archive files.
|
||||||
|
pub files: Vec<String>,
|
||||||
|
/// The owner of group
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub owner: Option<Authid>,
|
||||||
|
/// The first line from group "notes"
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"backup-type": {
|
||||||
|
schema: BACKUP_TYPE_SCHEMA,
|
||||||
|
},
|
||||||
|
"backup-id": {
|
||||||
|
schema: BACKUP_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
"backup-time": {
|
||||||
|
schema: BACKUP_TIME_SCHEMA,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Prune result.
|
||||||
|
pub struct PruneListItem {
|
||||||
|
pub backup_type: String, // enum
|
||||||
|
pub backup_id: String,
|
||||||
|
pub backup_time: i64,
|
||||||
|
/// Keep snapshot
|
||||||
|
pub keep: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
ct: {
|
||||||
|
type: TypeCounts,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
host: {
|
||||||
|
type: TypeCounts,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
vm: {
|
||||||
|
type: TypeCounts,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
other: {
|
||||||
|
type: TypeCounts,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize, Default)]
|
||||||
|
/// Counts of groups/snapshots per BackupType.
|
||||||
|
pub struct Counts {
|
||||||
|
/// The counts for CT backups
|
||||||
|
pub ct: Option<TypeCounts>,
|
||||||
|
/// The counts for Host backups
|
||||||
|
pub host: Option<TypeCounts>,
|
||||||
|
/// The counts for VM backups
|
||||||
|
pub vm: Option<TypeCounts>,
|
||||||
|
/// The counts for other backup types
|
||||||
|
pub other: Option<TypeCounts>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Serialize, Deserialize, Default)]
|
||||||
|
/// Backup Type group/snapshot counts.
|
||||||
|
pub struct TypeCounts {
|
||||||
|
/// The number of groups of the type.
|
||||||
|
pub groups: u64,
|
||||||
|
/// The number of snapshots of the type.
|
||||||
|
pub snapshots: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"upid": {
|
||||||
|
optional: true,
|
||||||
|
type: UPID,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Clone, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Garbage collection status.
|
||||||
|
pub struct GarbageCollectionStatus {
|
||||||
|
pub upid: Option<String>,
|
||||||
|
/// Number of processed index files.
|
||||||
|
pub index_file_count: usize,
|
||||||
|
/// Sum of bytes referred by index files.
|
||||||
|
pub index_data_bytes: u64,
|
||||||
|
/// Bytes used on disk.
|
||||||
|
pub disk_bytes: u64,
|
||||||
|
/// Chunks used on disk.
|
||||||
|
pub disk_chunks: usize,
|
||||||
|
/// Sum of removed bytes.
|
||||||
|
pub removed_bytes: u64,
|
||||||
|
/// Number of removed chunks.
|
||||||
|
pub removed_chunks: usize,
|
||||||
|
/// Sum of pending bytes (pending removal - kept for safety).
|
||||||
|
pub pending_bytes: u64,
|
||||||
|
/// Number of pending chunks (pending removal - kept for safety).
|
||||||
|
pub pending_chunks: usize,
|
||||||
|
/// Number of chunks marked as .bad by verify that have been removed by GC.
|
||||||
|
pub removed_bad: usize,
|
||||||
|
/// Number of chunks still marked as .bad after garbage collection.
|
||||||
|
pub still_bad: usize,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Default for GarbageCollectionStatus {
|
||||||
|
fn default() -> Self {
|
||||||
|
GarbageCollectionStatus {
|
||||||
|
upid: None,
|
||||||
|
index_file_count: 0,
|
||||||
|
index_data_bytes: 0,
|
||||||
|
disk_bytes: 0,
|
||||||
|
disk_chunks: 0,
|
||||||
|
removed_bytes: 0,
|
||||||
|
removed_chunks: 0,
|
||||||
|
pending_bytes: 0,
|
||||||
|
pending_chunks: 0,
|
||||||
|
removed_bad: 0,
|
||||||
|
still_bad: 0,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"gc-status": {
|
||||||
|
type: GarbageCollectionStatus,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
counts: {
|
||||||
|
type: Counts,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Overall Datastore status and useful information.
|
||||||
|
pub struct DataStoreStatus {
|
||||||
|
/// Total space (bytes).
|
||||||
|
pub total: u64,
|
||||||
|
/// Used space (bytes).
|
||||||
|
pub used: u64,
|
||||||
|
/// Available space (bytes).
|
||||||
|
pub avail: u64,
|
||||||
|
/// Status of last GC
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub gc_status: Option<GarbageCollectionStatus>,
|
||||||
|
/// Group/Snapshot counts
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub counts: Option<Counts>,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const ADMIN_DATASTORE_LIST_SNAPSHOTS_RETURN_TYPE: ReturnType = ReturnType {
|
||||||
|
optional: false,
|
||||||
|
schema: &ArraySchema::new(
|
||||||
|
"Returns the list of snapshots.",
|
||||||
|
&SnapshotListItem::API_SCHEMA,
|
||||||
|
).schema(),
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const ADMIN_DATASTORE_LIST_SNAPSHOT_FILES_RETURN_TYPE: ReturnType = ReturnType {
|
||||||
|
optional: false,
|
||||||
|
schema: &ArraySchema::new(
|
||||||
|
"Returns the list of archive files inside a backup snapshots.",
|
||||||
|
&BackupContent::API_SCHEMA,
|
||||||
|
).schema(),
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const ADMIN_DATASTORE_LIST_GROUPS_RETURN_TYPE: ReturnType = ReturnType {
|
||||||
|
optional: false,
|
||||||
|
schema: &ArraySchema::new(
|
||||||
|
"Returns the list of backup groups.",
|
||||||
|
&GroupListItem::API_SCHEMA,
|
||||||
|
).schema(),
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const ADMIN_DATASTORE_PRUNE_RETURN_TYPE: ReturnType = ReturnType {
|
||||||
|
optional: false,
|
||||||
|
schema: &ArraySchema::new(
|
||||||
|
"Returns the list of snapshots and a flag indicating if there are kept or removed.",
|
||||||
|
&PruneListItem::API_SCHEMA,
|
||||||
|
).schema(),
|
||||||
|
};
|
@ -1,7 +1,8 @@
|
|||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use proxmox::api::api;
|
|
||||||
|
|
||||||
#[api()]
|
use proxmox_schema::api;
|
||||||
|
|
||||||
|
#[api]
|
||||||
#[derive(Serialize, Deserialize)]
|
#[derive(Serialize, Deserialize)]
|
||||||
#[serde(rename_all = "kebab-case")]
|
#[serde(rename_all = "kebab-case")]
|
||||||
/// General status information about a running VM file-restore daemon
|
/// General status information about a running VM file-restore daemon
|
||||||
@ -12,4 +13,3 @@ pub struct RestoreDaemonStatus {
|
|||||||
/// not set, as then the status call will have reset the timer before returning the value
|
/// not set, as then the status call will have reset the timer before returning the value
|
||||||
pub timeout: i64,
|
pub timeout: i64,
|
||||||
}
|
}
|
||||||
|
|
341
pbs-api-types/src/human_byte.rs
Normal file
341
pbs-api-types/src/human_byte.rs
Normal file
@ -0,0 +1,341 @@
|
|||||||
|
use anyhow::{bail, Error};
|
||||||
|
|
||||||
|
use proxmox_schema::{ApiStringFormat, ApiType, Schema, StringSchema, UpdaterType};
|
||||||
|
|
||||||
|
/// Size units for byte sizes
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq)]
|
||||||
|
pub enum SizeUnit {
|
||||||
|
Byte,
|
||||||
|
// SI (base 10)
|
||||||
|
KByte,
|
||||||
|
MByte,
|
||||||
|
GByte,
|
||||||
|
TByte,
|
||||||
|
PByte,
|
||||||
|
// IEC (base 2)
|
||||||
|
Kibi,
|
||||||
|
Mebi,
|
||||||
|
Gibi,
|
||||||
|
Tebi,
|
||||||
|
Pebi,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl SizeUnit {
|
||||||
|
/// Returns the scaling factor
|
||||||
|
pub fn factor(&self) -> f64 {
|
||||||
|
match self {
|
||||||
|
SizeUnit::Byte => 1.0,
|
||||||
|
// SI (base 10)
|
||||||
|
SizeUnit::KByte => 1_000.0,
|
||||||
|
SizeUnit::MByte => 1_000_000.0,
|
||||||
|
SizeUnit::GByte => 1_000_000_000.0,
|
||||||
|
SizeUnit::TByte => 1_000_000_000_000.0,
|
||||||
|
SizeUnit::PByte => 1_000_000_000_000_000.0,
|
||||||
|
// IEC (base 2)
|
||||||
|
SizeUnit::Kibi => 1024.0,
|
||||||
|
SizeUnit::Mebi => 1024.0 * 1024.0,
|
||||||
|
SizeUnit::Gibi => 1024.0 * 1024.0 * 1024.0,
|
||||||
|
SizeUnit::Tebi => 1024.0 * 1024.0 * 1024.0 * 1024.0,
|
||||||
|
SizeUnit::Pebi => 1024.0 * 1024.0 * 1024.0 * 1024.0 * 1024.0,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// gets the biggest possible unit still having a value greater zero before the decimal point
|
||||||
|
/// 'binary' specifies if IEC (base 2) units should be used or SI (base 10) ones
|
||||||
|
pub fn auto_scale(size: f64, binary: bool) -> SizeUnit {
|
||||||
|
if binary {
|
||||||
|
let bits = 64 - (size as u64).leading_zeros();
|
||||||
|
match bits {
|
||||||
|
51.. => SizeUnit::Pebi,
|
||||||
|
41..=50 => SizeUnit::Tebi,
|
||||||
|
31..=40 => SizeUnit::Gibi,
|
||||||
|
21..=30 => SizeUnit::Mebi,
|
||||||
|
11..=20 => SizeUnit::Kibi,
|
||||||
|
_ => SizeUnit::Byte,
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if size >= 1_000_000_000_000_000.0 {
|
||||||
|
SizeUnit::PByte
|
||||||
|
} else if size >= 1_000_000_000_000.0 {
|
||||||
|
SizeUnit::TByte
|
||||||
|
} else if size >= 1_000_000_000.0 {
|
||||||
|
SizeUnit::GByte
|
||||||
|
} else if size >= 1_000_000.0 {
|
||||||
|
SizeUnit::MByte
|
||||||
|
} else if size >= 1_000.0 {
|
||||||
|
SizeUnit::KByte
|
||||||
|
} else {
|
||||||
|
SizeUnit::Byte
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Returns the string repesentation
|
||||||
|
impl std::fmt::Display for SizeUnit {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
|
match self {
|
||||||
|
SizeUnit::Byte => write!(f, "B"),
|
||||||
|
// SI (base 10)
|
||||||
|
SizeUnit::KByte => write!(f, "KB"),
|
||||||
|
SizeUnit::MByte => write!(f, "MB"),
|
||||||
|
SizeUnit::GByte => write!(f, "GB"),
|
||||||
|
SizeUnit::TByte => write!(f, "TB"),
|
||||||
|
SizeUnit::PByte => write!(f, "PB"),
|
||||||
|
// IEC (base 2)
|
||||||
|
SizeUnit::Kibi => write!(f, "KiB"),
|
||||||
|
SizeUnit::Mebi => write!(f, "MiB"),
|
||||||
|
SizeUnit::Gibi => write!(f, "GiB"),
|
||||||
|
SizeUnit::Tebi => write!(f, "TiB"),
|
||||||
|
SizeUnit::Pebi => write!(f, "PiB"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Strips a trailing SizeUnit inclusive trailing whitespace
|
||||||
|
/// Supports both IEC and SI based scales, the B/b byte symbol is optional.
|
||||||
|
fn strip_unit(v: &str) -> (&str, SizeUnit) {
|
||||||
|
let v = v.strip_suffix(&['b', 'B'][..]).unwrap_or(v); // byte is implied anyway
|
||||||
|
|
||||||
|
let (v, binary) = match v.strip_suffix('i') {
|
||||||
|
Some(n) => (n, true),
|
||||||
|
None => (v, false),
|
||||||
|
};
|
||||||
|
|
||||||
|
let mut unit = SizeUnit::Byte;
|
||||||
|
(v.strip_suffix(|c: char| match c {
|
||||||
|
'k' | 'K' if !binary => { unit = SizeUnit::KByte; true }
|
||||||
|
'm' | 'M' if !binary => { unit = SizeUnit::MByte; true }
|
||||||
|
'g' | 'G' if !binary => { unit = SizeUnit::GByte; true }
|
||||||
|
't' | 'T' if !binary => { unit = SizeUnit::TByte; true }
|
||||||
|
'p' | 'P' if !binary => { unit = SizeUnit::PByte; true }
|
||||||
|
// binary (IEC recommended) variants
|
||||||
|
'k' | 'K' if binary => { unit = SizeUnit::Kibi; true }
|
||||||
|
'm' | 'M' if binary => { unit = SizeUnit::Mebi; true }
|
||||||
|
'g' | 'G' if binary => { unit = SizeUnit::Gibi; true }
|
||||||
|
't' | 'T' if binary => { unit = SizeUnit::Tebi; true }
|
||||||
|
'p' | 'P' if binary => { unit = SizeUnit::Pebi; true }
|
||||||
|
_ => false
|
||||||
|
}).unwrap_or(v).trim_end(), unit)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Byte size which can be displayed in a human friendly way
|
||||||
|
#[derive(Debug, Copy, Clone, UpdaterType)]
|
||||||
|
pub struct HumanByte {
|
||||||
|
/// The siginficant value, it does not includes any factor of the `unit`
|
||||||
|
size: f64,
|
||||||
|
/// The scale/unit of the value
|
||||||
|
unit: SizeUnit,
|
||||||
|
}
|
||||||
|
|
||||||
|
fn verify_human_byte(s: &str) -> Result<(), Error> {
|
||||||
|
match s.parse::<HumanByte>() {
|
||||||
|
Ok(_) => Ok(()),
|
||||||
|
Err(err) => bail!("byte-size parse error for '{}': {}", s, err),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl ApiType for HumanByte {
|
||||||
|
const API_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Byte size with optional unit (B, KB (base 10), MB, GB, ..., KiB (base 2), MiB, Gib, ...).",
|
||||||
|
)
|
||||||
|
.format(&ApiStringFormat::VerifyFn(verify_human_byte))
|
||||||
|
.min_length(1)
|
||||||
|
.max_length(64)
|
||||||
|
.schema();
|
||||||
|
}
|
||||||
|
|
||||||
|
impl HumanByte {
|
||||||
|
/// Create instance with size and unit (size must be positive)
|
||||||
|
pub fn with_unit(size: f64, unit: SizeUnit) -> Result<Self, Error> {
|
||||||
|
if size < 0.0 {
|
||||||
|
bail!("byte size may not be negative");
|
||||||
|
}
|
||||||
|
Ok(HumanByte { size, unit })
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a new instance with optimal binary unit computed
|
||||||
|
pub fn new_binary(size: f64) -> Self {
|
||||||
|
let unit = SizeUnit::auto_scale(size, true);
|
||||||
|
HumanByte { size: size / unit.factor(), unit }
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a new instance with optimal decimal unit computed
|
||||||
|
pub fn new_decimal(size: f64) -> Self {
|
||||||
|
let unit = SizeUnit::auto_scale(size, false);
|
||||||
|
HumanByte { size: size / unit.factor(), unit }
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Returns the size as u64 number of bytes
|
||||||
|
pub fn as_u64(&self) -> u64 {
|
||||||
|
self.as_f64() as u64
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Returns the size as f64 number of bytes
|
||||||
|
pub fn as_f64(&self) -> f64 {
|
||||||
|
self.size * self.unit.factor()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Returns a copy with optimal binary unit computed
|
||||||
|
pub fn auto_scale_binary(self) -> Self {
|
||||||
|
HumanByte::new_binary(self.as_f64())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Returns a copy with optimal decimal unit computed
|
||||||
|
pub fn auto_scale_decimal(self) -> Self {
|
||||||
|
HumanByte::new_decimal(self.as_f64())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<u64> for HumanByte {
|
||||||
|
fn from(v: u64) -> Self {
|
||||||
|
HumanByte::new_binary(v as f64)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl From<usize> for HumanByte {
|
||||||
|
fn from(v: usize) -> Self {
|
||||||
|
HumanByte::new_binary(v as f64)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::fmt::Display for HumanByte {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
|
let precision = f.precision().unwrap_or(3) as f64;
|
||||||
|
let precision_factor = 1.0 * 10.0_f64.powf(precision);
|
||||||
|
// this could cause loss of information, rust has sadly no shortest-max-X flt2dec fmt yet
|
||||||
|
let size = ((self.size * precision_factor).round()) / precision_factor;
|
||||||
|
write!(f, "{} {}", size, self.unit)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::str::FromStr for HumanByte {
|
||||||
|
type Err = Error;
|
||||||
|
|
||||||
|
fn from_str(v: &str) -> Result<Self, Error> {
|
||||||
|
let (v, unit) = strip_unit(v);
|
||||||
|
HumanByte::with_unit(v.parse()?, unit)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
proxmox::forward_deserialize_to_from_str!(HumanByte);
|
||||||
|
proxmox::forward_serialize_to_display!(HumanByte);
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_human_byte_parser() -> Result<(), Error> {
|
||||||
|
assert!("-10".parse::<HumanByte>().is_err()); // negative size
|
||||||
|
|
||||||
|
fn do_test(v: &str, size: f64, unit: SizeUnit, as_str: &str) -> Result<(), Error> {
|
||||||
|
let h: HumanByte = v.parse()?;
|
||||||
|
|
||||||
|
if h.size != size {
|
||||||
|
bail!("got unexpected size for '{}' ({} != {})", v, h.size, size);
|
||||||
|
}
|
||||||
|
if h.unit != unit {
|
||||||
|
bail!("got unexpected unit for '{}' ({:?} != {:?})", v, h.unit, unit);
|
||||||
|
}
|
||||||
|
|
||||||
|
let new = h.to_string();
|
||||||
|
if &new != as_str {
|
||||||
|
bail!("to_string failed for '{}' ({:?} != {:?})", v, new, as_str);
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
fn test(v: &str, size: f64, unit: SizeUnit, as_str: &str) -> bool {
|
||||||
|
match do_test(v, size, unit, as_str) {
|
||||||
|
Ok(_) => true,
|
||||||
|
Err(err) => {
|
||||||
|
eprintln!("{}", err); // makes debugging easier
|
||||||
|
false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
assert!(test("14", 14.0, SizeUnit::Byte, "14 B"));
|
||||||
|
assert!(test("14.4", 14.4, SizeUnit::Byte, "14.4 B"));
|
||||||
|
assert!(test("14.45", 14.45, SizeUnit::Byte, "14.45 B"));
|
||||||
|
assert!(test("14.456", 14.456, SizeUnit::Byte, "14.456 B"));
|
||||||
|
assert!(test("14.4567", 14.4567, SizeUnit::Byte, "14.457 B"));
|
||||||
|
|
||||||
|
let h: HumanByte = "1.2345678".parse()?;
|
||||||
|
assert_eq!(&format!("{:.0}", h), "1 B");
|
||||||
|
assert_eq!(&format!("{:.0}", h.as_f64()), "1"); // use as_f64 to get raw bytes without unit
|
||||||
|
assert_eq!(&format!("{:.1}", h), "1.2 B");
|
||||||
|
assert_eq!(&format!("{:.2}", h), "1.23 B");
|
||||||
|
assert_eq!(&format!("{:.3}", h), "1.235 B");
|
||||||
|
assert_eq!(&format!("{:.4}", h), "1.2346 B");
|
||||||
|
assert_eq!(&format!("{:.5}", h), "1.23457 B");
|
||||||
|
assert_eq!(&format!("{:.6}", h), "1.234568 B");
|
||||||
|
assert_eq!(&format!("{:.7}", h), "1.2345678 B");
|
||||||
|
assert_eq!(&format!("{:.8}", h), "1.2345678 B");
|
||||||
|
|
||||||
|
assert!(test("987654321", 987654321.0, SizeUnit::Byte, "987654321 B"));
|
||||||
|
|
||||||
|
assert!(test("1300b", 1300.0, SizeUnit::Byte, "1300 B"));
|
||||||
|
assert!(test("1300B", 1300.0, SizeUnit::Byte, "1300 B"));
|
||||||
|
assert!(test("1300 B", 1300.0, SizeUnit::Byte, "1300 B"));
|
||||||
|
assert!(test("1300 b", 1300.0, SizeUnit::Byte, "1300 B"));
|
||||||
|
|
||||||
|
assert!(test("1.5KB", 1.5, SizeUnit::KByte, "1.5 KB"));
|
||||||
|
assert!(test("1.5kb", 1.5, SizeUnit::KByte, "1.5 KB"));
|
||||||
|
assert!(test("1.654321MB", 1.654_321, SizeUnit::MByte, "1.654 MB"));
|
||||||
|
|
||||||
|
assert!(test("2.0GB", 2.0, SizeUnit::GByte, "2 GB"));
|
||||||
|
|
||||||
|
assert!(test("1.4TB", 1.4, SizeUnit::TByte, "1.4 TB"));
|
||||||
|
assert!(test("1.4tb", 1.4, SizeUnit::TByte, "1.4 TB"));
|
||||||
|
|
||||||
|
assert!(test("2KiB", 2.0, SizeUnit::Kibi, "2 KiB"));
|
||||||
|
assert!(test("2Ki", 2.0, SizeUnit::Kibi, "2 KiB"));
|
||||||
|
assert!(test("2kib", 2.0, SizeUnit::Kibi, "2 KiB"));
|
||||||
|
|
||||||
|
assert!(test("2.3454MiB", 2.3454, SizeUnit::Mebi, "2.345 MiB"));
|
||||||
|
assert!(test("2.3456MiB", 2.3456, SizeUnit::Mebi, "2.346 MiB"));
|
||||||
|
|
||||||
|
assert!(test("4gib", 4.0, SizeUnit::Gibi, "4 GiB"));
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_human_byte_auto_unit_decimal() {
|
||||||
|
fn convert(b: u64) -> String {
|
||||||
|
HumanByte::new_decimal(b as f64).to_string()
|
||||||
|
}
|
||||||
|
assert_eq!(convert(987), "987 B");
|
||||||
|
assert_eq!(convert(1022), "1.022 KB");
|
||||||
|
assert_eq!(convert(9_000), "9 KB");
|
||||||
|
assert_eq!(convert(1_000), "1 KB");
|
||||||
|
assert_eq!(convert(1_000_000), "1 MB");
|
||||||
|
assert_eq!(convert(1_000_000_000), "1 GB");
|
||||||
|
assert_eq!(convert(1_000_000_000_000), "1 TB");
|
||||||
|
assert_eq!(convert(1_000_000_000_000_000), "1 PB");
|
||||||
|
|
||||||
|
assert_eq!(convert((1 << 30) + 103 * (1 << 20)), "1.182 GB");
|
||||||
|
assert_eq!(convert((1 << 30) + 128 * (1 << 20)), "1.208 GB");
|
||||||
|
assert_eq!(convert((2 << 50) + 500 * (1 << 40)), "2.802 PB");
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_human_byte_auto_unit_binary() {
|
||||||
|
fn convert(b: u64) -> String {
|
||||||
|
HumanByte::from(b).to_string()
|
||||||
|
}
|
||||||
|
assert_eq!(convert(0), "0 B");
|
||||||
|
assert_eq!(convert(987), "987 B");
|
||||||
|
assert_eq!(convert(1022), "1022 B");
|
||||||
|
assert_eq!(convert(9_000), "8.789 KiB");
|
||||||
|
assert_eq!(convert(10_000_000), "9.537 MiB");
|
||||||
|
assert_eq!(convert(10_000_000_000), "9.313 GiB");
|
||||||
|
assert_eq!(convert(10_000_000_000_000), "9.095 TiB");
|
||||||
|
|
||||||
|
assert_eq!(convert(1 << 10), "1 KiB");
|
||||||
|
assert_eq!(convert((1 << 10) * 10), "10 KiB");
|
||||||
|
assert_eq!(convert(1 << 20), "1 MiB");
|
||||||
|
assert_eq!(convert(1 << 30), "1 GiB");
|
||||||
|
assert_eq!(convert(1 << 40), "1 TiB");
|
||||||
|
assert_eq!(convert(1 << 50), "1 PiB");
|
||||||
|
|
||||||
|
assert_eq!(convert((1 << 30) + 103 * (1 << 20)), "1.101 GiB");
|
||||||
|
assert_eq!(convert((1 << 30) + 128 * (1 << 20)), "1.125 GiB");
|
||||||
|
assert_eq!(convert((1 << 40) + 128 * (1 << 30)), "1.125 TiB");
|
||||||
|
assert_eq!(convert((2 << 50) + 512 * (1 << 40)), "2.5 PiB");
|
||||||
|
}
|
464
pbs-api-types/src/jobs.rs
Normal file
464
pbs-api-types/src/jobs.rs
Normal file
@ -0,0 +1,464 @@
|
|||||||
|
use anyhow::format_err;
|
||||||
|
use std::str::FromStr;
|
||||||
|
|
||||||
|
use regex::Regex;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::*;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
Userid, Authid, RateLimitConfig,
|
||||||
|
REMOTE_ID_SCHEMA, DRIVE_NAME_SCHEMA, MEDIA_POOL_NAME_SCHEMA,
|
||||||
|
SINGLE_LINE_COMMENT_SCHEMA, PROXMOX_SAFE_ID_FORMAT, DATASTORE_SCHEMA,
|
||||||
|
BACKUP_GROUP_SCHEMA, BACKUP_TYPE_SCHEMA,
|
||||||
|
};
|
||||||
|
|
||||||
|
const_regex!{
|
||||||
|
|
||||||
|
/// Regex for verification jobs 'DATASTORE:ACTUAL_JOB_ID'
|
||||||
|
pub VERIFICATION_JOB_WORKER_ID_REGEX = concat!(r"^(", PROXMOX_SAFE_ID_REGEX_STR!(), r"):");
|
||||||
|
/// Regex for sync jobs 'REMOTE:REMOTE_DATASTORE:LOCAL_DATASTORE:ACTUAL_JOB_ID'
|
||||||
|
pub SYNC_JOB_WORKER_ID_REGEX = concat!(r"^(", PROXMOX_SAFE_ID_REGEX_STR!(), r"):(", PROXMOX_SAFE_ID_REGEX_STR!(), r"):(", PROXMOX_SAFE_ID_REGEX_STR!(), r"):");
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const JOB_ID_SCHEMA: Schema = StringSchema::new("Job ID.")
|
||||||
|
.format(&PROXMOX_SAFE_ID_FORMAT)
|
||||||
|
.min_length(3)
|
||||||
|
.max_length(32)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const SYNC_SCHEDULE_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Run sync job at specified schedule.")
|
||||||
|
.format(&ApiStringFormat::VerifyFn(proxmox_time::verify_calendar_event))
|
||||||
|
.type_text("<calendar-event>")
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const GC_SCHEDULE_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Run garbage collection job at specified schedule.")
|
||||||
|
.format(&ApiStringFormat::VerifyFn(proxmox_time::verify_calendar_event))
|
||||||
|
.type_text("<calendar-event>")
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PRUNE_SCHEDULE_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Run prune job at specified schedule.")
|
||||||
|
.format(&ApiStringFormat::VerifyFn(proxmox_time::verify_calendar_event))
|
||||||
|
.type_text("<calendar-event>")
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const VERIFICATION_SCHEDULE_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Run verify job at specified schedule.")
|
||||||
|
.format(&ApiStringFormat::VerifyFn(proxmox_time::verify_calendar_event))
|
||||||
|
.type_text("<calendar-event>")
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const REMOVE_VANISHED_BACKUPS_SCHEMA: Schema = BooleanSchema::new(
|
||||||
|
"Delete vanished backups. This remove the local copy if the remote backup was deleted.")
|
||||||
|
.default(false)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"next-run": {
|
||||||
|
description: "Estimated time of the next run (UNIX epoch).",
|
||||||
|
optional: true,
|
||||||
|
type: Integer,
|
||||||
|
},
|
||||||
|
"last-run-state": {
|
||||||
|
description: "Result of the last run.",
|
||||||
|
optional: true,
|
||||||
|
type: String,
|
||||||
|
},
|
||||||
|
"last-run-upid": {
|
||||||
|
description: "Task UPID of the last run.",
|
||||||
|
optional: true,
|
||||||
|
type: String,
|
||||||
|
},
|
||||||
|
"last-run-endtime": {
|
||||||
|
description: "Endtime of the last run.",
|
||||||
|
optional: true,
|
||||||
|
type: Integer,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize,Default)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Job Scheduling Status
|
||||||
|
pub struct JobScheduleStatus {
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub next_run: Option<i64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub last_run_state: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub last_run_upid: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub last_run_endtime: Option<i64>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
/// When do we send notifications
|
||||||
|
pub enum Notify {
|
||||||
|
/// Never send notification
|
||||||
|
Never,
|
||||||
|
/// Send notifications for failed and successful jobs
|
||||||
|
Always,
|
||||||
|
/// Send notifications for failed jobs only
|
||||||
|
Error,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
gc: {
|
||||||
|
type: Notify,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
verify: {
|
||||||
|
type: Notify,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
sync: {
|
||||||
|
type: Notify,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
|
/// Datastore notify settings
|
||||||
|
pub struct DatastoreNotify {
|
||||||
|
/// Garbage collection settings
|
||||||
|
pub gc: Option<Notify>,
|
||||||
|
/// Verify job setting
|
||||||
|
pub verify: Option<Notify>,
|
||||||
|
/// Sync job setting
|
||||||
|
pub sync: Option<Notify>,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const DATASTORE_NOTIFY_STRING_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Datastore notification setting")
|
||||||
|
.format(&ApiStringFormat::PropertyString(&DatastoreNotify::API_SCHEMA))
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const IGNORE_VERIFIED_BACKUPS_SCHEMA: Schema = BooleanSchema::new(
|
||||||
|
"Do not verify backups that are already verified if their verification is not outdated.")
|
||||||
|
.default(true)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const VERIFICATION_OUTDATED_AFTER_SCHEMA: Schema = IntegerSchema::new(
|
||||||
|
"Days after that a verification becomes outdated")
|
||||||
|
.minimum(1)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
id: {
|
||||||
|
schema: JOB_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
store: {
|
||||||
|
schema: DATASTORE_SCHEMA,
|
||||||
|
},
|
||||||
|
"ignore-verified": {
|
||||||
|
optional: true,
|
||||||
|
schema: IGNORE_VERIFIED_BACKUPS_SCHEMA,
|
||||||
|
},
|
||||||
|
"outdated-after": {
|
||||||
|
optional: true,
|
||||||
|
schema: VERIFICATION_OUTDATED_AFTER_SCHEMA,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
schedule: {
|
||||||
|
optional: true,
|
||||||
|
schema: VERIFICATION_SCHEDULE_SCHEMA,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize,Updater)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Verification Job
|
||||||
|
pub struct VerificationJobConfig {
|
||||||
|
/// unique ID to address this job
|
||||||
|
#[updater(skip)]
|
||||||
|
pub id: String,
|
||||||
|
/// the datastore ID this verificaiton job affects
|
||||||
|
pub store: String,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
/// if not set to false, check the age of the last snapshot verification to filter
|
||||||
|
/// out recent ones, depending on 'outdated_after' configuration.
|
||||||
|
pub ignore_verified: Option<bool>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
/// Reverify snapshots after X days, never if 0. Ignored if 'ignore_verified' is false.
|
||||||
|
pub outdated_after: Option<i64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
/// when to schedule this job in calendar event notation
|
||||||
|
pub schedule: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
config: {
|
||||||
|
type: VerificationJobConfig,
|
||||||
|
},
|
||||||
|
status: {
|
||||||
|
type: JobScheduleStatus,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Status of Verification Job
|
||||||
|
pub struct VerificationJobStatus {
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub config: VerificationJobConfig,
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub status: JobScheduleStatus,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
store: {
|
||||||
|
schema: DATASTORE_SCHEMA,
|
||||||
|
},
|
||||||
|
pool: {
|
||||||
|
schema: MEDIA_POOL_NAME_SCHEMA,
|
||||||
|
},
|
||||||
|
drive: {
|
||||||
|
schema: DRIVE_NAME_SCHEMA,
|
||||||
|
},
|
||||||
|
"eject-media": {
|
||||||
|
description: "Eject media upon job completion.",
|
||||||
|
type: bool,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"export-media-set": {
|
||||||
|
description: "Export media set upon job completion.",
|
||||||
|
type: bool,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"latest-only": {
|
||||||
|
description: "Backup latest snapshots only.",
|
||||||
|
type: bool,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"notify-user": {
|
||||||
|
optional: true,
|
||||||
|
type: Userid,
|
||||||
|
},
|
||||||
|
"group-filter": {
|
||||||
|
schema: GROUP_FILTER_LIST_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize,Clone,Updater)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Tape Backup Job Setup
|
||||||
|
pub struct TapeBackupJobSetup {
|
||||||
|
pub store: String,
|
||||||
|
pub pool: String,
|
||||||
|
pub drive: String,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub eject_media: Option<bool>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub export_media_set: Option<bool>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub latest_only: Option<bool>,
|
||||||
|
/// Send job email notification to this user
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub notify_user: Option<Userid>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub group_filter: Option<Vec<GroupFilter>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
id: {
|
||||||
|
schema: JOB_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
setup: {
|
||||||
|
type: TapeBackupJobSetup,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
schedule: {
|
||||||
|
optional: true,
|
||||||
|
schema: SYNC_SCHEDULE_SCHEMA,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize,Clone,Updater)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Tape Backup Job
|
||||||
|
pub struct TapeBackupJobConfig {
|
||||||
|
#[updater(skip)]
|
||||||
|
pub id: String,
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub setup: TapeBackupJobSetup,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub schedule: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
config: {
|
||||||
|
type: TapeBackupJobConfig,
|
||||||
|
},
|
||||||
|
status: {
|
||||||
|
type: JobScheduleStatus,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Status of Tape Backup Job
|
||||||
|
pub struct TapeBackupJobStatus {
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub config: TapeBackupJobConfig,
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub status: JobScheduleStatus,
|
||||||
|
/// Next tape used (best guess)
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub next_media_label: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Debug)]
|
||||||
|
/// Filter for matching `BackupGroup`s, for use with `BackupGroup::filter`.
|
||||||
|
pub enum GroupFilter {
|
||||||
|
/// BackupGroup type - either `vm`, `ct`, or `host`.
|
||||||
|
BackupType(String),
|
||||||
|
/// Full identifier of BackupGroup, including type
|
||||||
|
Group(String),
|
||||||
|
/// A regular expression matched against the full identifier of the BackupGroup
|
||||||
|
Regex(Regex),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::str::FromStr for GroupFilter {
|
||||||
|
type Err = anyhow::Error;
|
||||||
|
|
||||||
|
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||||
|
match s.split_once(":") {
|
||||||
|
Some(("group", value)) => parse_simple_value(value, &BACKUP_GROUP_SCHEMA).map(|_| GroupFilter::Group(value.to_string())),
|
||||||
|
Some(("type", value)) => parse_simple_value(value, &BACKUP_TYPE_SCHEMA).map(|_| GroupFilter::BackupType(value.to_string())),
|
||||||
|
Some(("regex", value)) => Ok(GroupFilter::Regex(Regex::new(value)?)),
|
||||||
|
Some((ty, _value)) => Err(format_err!("expected 'group', 'type' or 'regex' prefix, got '{}'", ty)),
|
||||||
|
None => Err(format_err!("input doesn't match expected format '<group:GROUP||type:<vm|ct|host>|regex:REGEX>'")),
|
||||||
|
}.map_err(|err| format_err!("'{}' - {}", s, err))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// used for serializing below, caution!
|
||||||
|
impl std::fmt::Display for GroupFilter {
|
||||||
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
|
match self {
|
||||||
|
GroupFilter::BackupType(backup_type) => write!(f, "type:{}", backup_type),
|
||||||
|
GroupFilter::Group(backup_group) => write!(f, "group:{}", backup_group),
|
||||||
|
GroupFilter::Regex(regex) => write!(f, "regex:{}", regex.as_str()),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
proxmox::forward_deserialize_to_from_str!(GroupFilter);
|
||||||
|
proxmox::forward_serialize_to_display!(GroupFilter);
|
||||||
|
|
||||||
|
fn verify_group_filter(input: &str) -> Result<(), anyhow::Error> {
|
||||||
|
GroupFilter::from_str(input).map(|_| ())
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const GROUP_FILTER_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Group filter based on group identifier ('group:GROUP'), group type ('type:<vm|ct|host>'), or regex ('regex:RE').")
|
||||||
|
.format(&ApiStringFormat::VerifyFn(verify_group_filter))
|
||||||
|
.type_text("<type:<vm|ct|host>|group:GROUP|regex:RE>")
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const GROUP_FILTER_LIST_SCHEMA: Schema = ArraySchema::new("List of group filters.", &GROUP_FILTER_SCHEMA).schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
id: {
|
||||||
|
schema: JOB_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
store: {
|
||||||
|
schema: DATASTORE_SCHEMA,
|
||||||
|
},
|
||||||
|
"owner": {
|
||||||
|
type: Authid,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
remote: {
|
||||||
|
schema: REMOTE_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
"remote-store": {
|
||||||
|
schema: DATASTORE_SCHEMA,
|
||||||
|
},
|
||||||
|
"remove-vanished": {
|
||||||
|
schema: REMOVE_VANISHED_BACKUPS_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
limit: {
|
||||||
|
type: RateLimitConfig,
|
||||||
|
},
|
||||||
|
schedule: {
|
||||||
|
optional: true,
|
||||||
|
schema: SYNC_SCHEDULE_SCHEMA,
|
||||||
|
},
|
||||||
|
"group-filter": {
|
||||||
|
schema: GROUP_FILTER_LIST_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize,Clone,Updater)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Sync Job
|
||||||
|
pub struct SyncJobConfig {
|
||||||
|
#[updater(skip)]
|
||||||
|
pub id: String,
|
||||||
|
pub store: String,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub owner: Option<Authid>,
|
||||||
|
pub remote: String,
|
||||||
|
pub remote_store: String,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub remove_vanished: Option<bool>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub schedule: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub group_filter: Option<Vec<GroupFilter>>,
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub limit: RateLimitConfig,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
config: {
|
||||||
|
type: SyncJobConfig,
|
||||||
|
},
|
||||||
|
status: {
|
||||||
|
type: JobScheduleStatus,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
|
||||||
|
#[derive(Serialize,Deserialize)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Status of Sync Job
|
||||||
|
pub struct SyncJobStatus {
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub config: SyncJobConfig,
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub status: JobScheduleStatus,
|
||||||
|
}
|
56
pbs-api-types/src/key_derivation.rs
Normal file
56
pbs-api-types/src/key_derivation.rs
Normal file
@ -0,0 +1,56 @@
|
|||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::api;
|
||||||
|
|
||||||
|
use crate::CERT_FINGERPRINT_SHA256_SCHEMA;
|
||||||
|
|
||||||
|
#[api(default: "scrypt")]
|
||||||
|
#[derive(Clone, Copy, Debug, Deserialize, Serialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
/// Key derivation function for password protected encryption keys.
|
||||||
|
pub enum Kdf {
|
||||||
|
/// Do not encrypt the key.
|
||||||
|
None,
|
||||||
|
/// Encrypt they key with a password using SCrypt.
|
||||||
|
Scrypt,
|
||||||
|
/// Encrtypt the Key with a password using PBKDF2
|
||||||
|
PBKDF2,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Default for Kdf {
|
||||||
|
#[inline]
|
||||||
|
fn default() -> Self {
|
||||||
|
Kdf::Scrypt
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
kdf: {
|
||||||
|
type: Kdf,
|
||||||
|
},
|
||||||
|
fingerprint: {
|
||||||
|
schema: CERT_FINGERPRINT_SHA256_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Deserialize, Serialize)]
|
||||||
|
/// Encryption Key Information
|
||||||
|
pub struct KeyInfo {
|
||||||
|
/// Path to key (if stored in a file)
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub path: Option<String>,
|
||||||
|
pub kdf: Kdf,
|
||||||
|
/// Key creation time
|
||||||
|
pub created: i64,
|
||||||
|
/// Key modification time
|
||||||
|
pub modified: i64,
|
||||||
|
/// Key fingerprint
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub fingerprint: Option<String>,
|
||||||
|
/// Password hint
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub hint: Option<String>,
|
||||||
|
}
|
||||||
|
|
468
pbs-api-types/src/lib.rs
Normal file
468
pbs-api-types/src/lib.rs
Normal file
@ -0,0 +1,468 @@
|
|||||||
|
//! Basic API types used by most of the PBS code.
|
||||||
|
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
use anyhow::bail;
|
||||||
|
|
||||||
|
use proxmox_schema::{
|
||||||
|
api, const_regex, ApiStringFormat, ApiType, ArraySchema, Schema, StringSchema, ReturnType,
|
||||||
|
};
|
||||||
|
use proxmox::{IPRE, IPRE_BRACKET, IPV4OCTET, IPV4RE, IPV6H16, IPV6LS32, IPV6RE};
|
||||||
|
use proxmox_time::parse_daily_duration;
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[macro_export]
|
||||||
|
macro_rules! PROXMOX_SAFE_ID_REGEX_STR { () => { r"(?:[A-Za-z0-9_][A-Za-z0-9._\-]*)" }; }
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[macro_export]
|
||||||
|
macro_rules! BACKUP_ID_RE { () => (r"[A-Za-z0-9_][A-Za-z0-9._\-]*") }
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[macro_export]
|
||||||
|
macro_rules! BACKUP_TYPE_RE { () => (r"(?:host|vm|ct)") }
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[macro_export]
|
||||||
|
macro_rules! BACKUP_TIME_RE { () => (r"[0-9]{4}-[0-9]{2}-[0-9]{2}T[0-9]{2}:[0-9]{2}:[0-9]{2}Z") }
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[macro_export]
|
||||||
|
macro_rules! SNAPSHOT_PATH_REGEX_STR {
|
||||||
|
() => (
|
||||||
|
concat!(r"(", BACKUP_TYPE_RE!(), ")/(", BACKUP_ID_RE!(), ")/(", BACKUP_TIME_RE!(), r")")
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
mod acl;
|
||||||
|
pub use acl::*;
|
||||||
|
|
||||||
|
mod datastore;
|
||||||
|
pub use datastore::*;
|
||||||
|
|
||||||
|
mod human_byte;
|
||||||
|
pub use human_byte::HumanByte;
|
||||||
|
|
||||||
|
mod jobs;
|
||||||
|
pub use jobs::*;
|
||||||
|
|
||||||
|
mod key_derivation;
|
||||||
|
pub use key_derivation::{Kdf, KeyInfo};
|
||||||
|
|
||||||
|
mod network;
|
||||||
|
pub use network::*;
|
||||||
|
|
||||||
|
#[macro_use]
|
||||||
|
mod userid;
|
||||||
|
pub use userid::Authid;
|
||||||
|
pub use userid::Userid;
|
||||||
|
pub use userid::{Realm, RealmRef};
|
||||||
|
pub use userid::{Tokenname, TokennameRef};
|
||||||
|
pub use userid::{Username, UsernameRef};
|
||||||
|
pub use userid::{PROXMOX_GROUP_ID_SCHEMA, PROXMOX_TOKEN_ID_SCHEMA, PROXMOX_TOKEN_NAME_SCHEMA};
|
||||||
|
|
||||||
|
#[macro_use]
|
||||||
|
mod user;
|
||||||
|
pub use user::*;
|
||||||
|
|
||||||
|
pub use proxmox_schema::upid::*;
|
||||||
|
|
||||||
|
mod crypto;
|
||||||
|
pub use crypto::{CryptMode, Fingerprint, bytes_as_fingerprint};
|
||||||
|
|
||||||
|
pub mod file_restore;
|
||||||
|
|
||||||
|
mod openid;
|
||||||
|
pub use openid::*;
|
||||||
|
|
||||||
|
mod remote;
|
||||||
|
pub use remote::*;
|
||||||
|
|
||||||
|
mod tape;
|
||||||
|
pub use tape::*;
|
||||||
|
|
||||||
|
mod traffic_control;
|
||||||
|
pub use traffic_control::*;
|
||||||
|
|
||||||
|
mod zfs;
|
||||||
|
pub use zfs::*;
|
||||||
|
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
#[macro_use]
|
||||||
|
mod local_macros {
|
||||||
|
macro_rules! DNS_LABEL { () => (r"(?:[a-zA-Z0-9](?:[a-zA-Z0-9\-]*[a-zA-Z0-9])?)") }
|
||||||
|
macro_rules! DNS_NAME { () => (concat!(r"(?:(?:", DNS_LABEL!() , r"\.)*", DNS_LABEL!(), ")")) }
|
||||||
|
macro_rules! CIDR_V4_REGEX_STR { () => (concat!(r"(?:", IPV4RE!(), r"/\d{1,2})$")) }
|
||||||
|
macro_rules! CIDR_V6_REGEX_STR { () => (concat!(r"(?:", IPV6RE!(), r"/\d{1,3})$")) }
|
||||||
|
macro_rules! DNS_ALIAS_LABEL { () => (r"(?:[a-zA-Z0-9_](?:[a-zA-Z0-9\-]*[a-zA-Z0-9])?)") }
|
||||||
|
macro_rules! DNS_ALIAS_NAME {
|
||||||
|
() => (concat!(r"(?:(?:", DNS_ALIAS_LABEL!() , r"\.)*", DNS_ALIAS_LABEL!(), ")"))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const_regex! {
|
||||||
|
pub IP_V4_REGEX = concat!(r"^", IPV4RE!(), r"$");
|
||||||
|
pub IP_V6_REGEX = concat!(r"^", IPV6RE!(), r"$");
|
||||||
|
pub IP_REGEX = concat!(r"^", IPRE!(), r"$");
|
||||||
|
pub CIDR_V4_REGEX = concat!(r"^", CIDR_V4_REGEX_STR!(), r"$");
|
||||||
|
pub CIDR_V6_REGEX = concat!(r"^", CIDR_V6_REGEX_STR!(), r"$");
|
||||||
|
pub CIDR_REGEX = concat!(r"^(?:", CIDR_V4_REGEX_STR!(), "|", CIDR_V6_REGEX_STR!(), r")$");
|
||||||
|
pub HOSTNAME_REGEX = r"^(?:[a-zA-Z0-9](?:[a-zA-Z0-9\-]*[a-zA-Z0-9])?)$";
|
||||||
|
pub DNS_NAME_REGEX = concat!(r"^", DNS_NAME!(), r"$");
|
||||||
|
pub DNS_ALIAS_REGEX = concat!(r"^", DNS_ALIAS_NAME!(), r"$");
|
||||||
|
pub DNS_NAME_OR_IP_REGEX = concat!(r"^(?:", DNS_NAME!(), "|", IPRE!(), r")$");
|
||||||
|
|
||||||
|
pub SHA256_HEX_REGEX = r"^[a-f0-9]{64}$"; // fixme: define in common_regex ?
|
||||||
|
|
||||||
|
pub PASSWORD_REGEX = r"^[[:^cntrl:]]*$"; // everything but control characters
|
||||||
|
|
||||||
|
pub UUID_REGEX = r"^[0-9a-f]{8}(?:-[0-9a-f]{4}){3}-[0-9a-f]{12}$";
|
||||||
|
|
||||||
|
pub SYSTEMD_DATETIME_REGEX = r"^\d{4}-\d{2}-\d{2}( \d{2}:\d{2}(:\d{2})?)?$"; // fixme: define in common_regex ?
|
||||||
|
|
||||||
|
pub FINGERPRINT_SHA256_REGEX = r"^(?:[0-9a-fA-F][0-9a-fA-F])(?::[0-9a-fA-F][0-9a-fA-F]){31}$";
|
||||||
|
|
||||||
|
/// Regex for safe identifiers.
|
||||||
|
///
|
||||||
|
/// This
|
||||||
|
/// [article](https://dwheeler.com/essays/fixing-unix-linux-filenames.html)
|
||||||
|
/// contains further information why it is reasonable to restict
|
||||||
|
/// names this way. This is not only useful for filenames, but for
|
||||||
|
/// any identifier command line tools work with.
|
||||||
|
pub PROXMOX_SAFE_ID_REGEX = concat!(r"^", PROXMOX_SAFE_ID_REGEX_STR!(), r"$");
|
||||||
|
|
||||||
|
pub SINGLE_LINE_COMMENT_REGEX = r"^[[:^cntrl:]]*$";
|
||||||
|
|
||||||
|
pub BACKUP_REPO_URL_REGEX = concat!(
|
||||||
|
r"^^(?:(?:(",
|
||||||
|
USER_ID_REGEX_STR!(), "|", APITOKEN_ID_REGEX_STR!(),
|
||||||
|
")@)?(",
|
||||||
|
DNS_NAME!(), "|", IPRE_BRACKET!(),
|
||||||
|
"):)?(?:([0-9]{1,5}):)?(", PROXMOX_SAFE_ID_REGEX_STR!(), r")$"
|
||||||
|
);
|
||||||
|
|
||||||
|
pub BLOCKDEVICE_NAME_REGEX = r"^(:?(:?h|s|x?v)d[a-z]+)|(:?nvme\d+n\d+)$";
|
||||||
|
pub SUBSCRIPTION_KEY_REGEX = concat!(r"^pbs(?:[cbsp])-[0-9a-f]{10}$");
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const IP_V4_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&IP_V4_REGEX);
|
||||||
|
pub const IP_V6_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&IP_V6_REGEX);
|
||||||
|
pub const IP_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&IP_REGEX);
|
||||||
|
pub const CIDR_V4_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&CIDR_V4_REGEX);
|
||||||
|
pub const CIDR_V6_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&CIDR_V6_REGEX);
|
||||||
|
pub const CIDR_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&CIDR_REGEX);
|
||||||
|
pub const PVE_CONFIG_DIGEST_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&SHA256_HEX_REGEX);
|
||||||
|
pub const PASSWORD_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&PASSWORD_REGEX);
|
||||||
|
pub const UUID_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&UUID_REGEX);
|
||||||
|
pub const BLOCKDEVICE_NAME_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&BLOCKDEVICE_NAME_REGEX);
|
||||||
|
pub const SUBSCRIPTION_KEY_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&SUBSCRIPTION_KEY_REGEX);
|
||||||
|
pub const SYSTEMD_DATETIME_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&SYSTEMD_DATETIME_REGEX);
|
||||||
|
pub const HOSTNAME_FORMAT: ApiStringFormat = ApiStringFormat::Pattern(&HOSTNAME_REGEX);
|
||||||
|
|
||||||
|
pub const DNS_ALIAS_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&DNS_ALIAS_REGEX);
|
||||||
|
|
||||||
|
pub const DAILY_DURATION_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::VerifyFn(|s| parse_daily_duration(s).map(drop));
|
||||||
|
|
||||||
|
pub const SEARCH_DOMAIN_SCHEMA: Schema =
|
||||||
|
StringSchema::new("Search domain for host-name lookup.").schema();
|
||||||
|
|
||||||
|
pub const FIRST_DNS_SERVER_SCHEMA: Schema =
|
||||||
|
StringSchema::new("First name server IP address.")
|
||||||
|
.format(&IP_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const SECOND_DNS_SERVER_SCHEMA: Schema =
|
||||||
|
StringSchema::new("Second name server IP address.")
|
||||||
|
.format(&IP_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const THIRD_DNS_SERVER_SCHEMA: Schema =
|
||||||
|
StringSchema::new("Third name server IP address.")
|
||||||
|
.format(&IP_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const HOSTNAME_SCHEMA: Schema = StringSchema::new("Hostname (as defined in RFC1123).")
|
||||||
|
.format(&HOSTNAME_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const DNS_NAME_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&DNS_NAME_REGEX);
|
||||||
|
|
||||||
|
pub const DNS_NAME_OR_IP_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&DNS_NAME_OR_IP_REGEX);
|
||||||
|
|
||||||
|
pub const DNS_NAME_OR_IP_SCHEMA: Schema = StringSchema::new("DNS name or IP address.")
|
||||||
|
.format(&DNS_NAME_OR_IP_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const NODE_SCHEMA: Schema = StringSchema::new("Node name (or 'localhost')")
|
||||||
|
.format(&ApiStringFormat::VerifyFn(|node| {
|
||||||
|
if node == "localhost" || node == proxmox::tools::nodename() {
|
||||||
|
Ok(())
|
||||||
|
} else {
|
||||||
|
bail!("no such node '{}'", node);
|
||||||
|
}
|
||||||
|
}))
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const TIME_ZONE_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Time zone. The file '/usr/share/zoneinfo/zone.tab' contains the list of valid names.")
|
||||||
|
.format(&SINGLE_LINE_COMMENT_FORMAT)
|
||||||
|
.min_length(2)
|
||||||
|
.max_length(64)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const BLOCKDEVICE_NAME_SCHEMA: Schema = StringSchema::new("Block device name (/sys/block/<name>).")
|
||||||
|
.format(&BLOCKDEVICE_NAME_FORMAT)
|
||||||
|
.min_length(3)
|
||||||
|
.max_length(64)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const DISK_ARRAY_SCHEMA: Schema = ArraySchema::new(
|
||||||
|
"Disk name list.", &BLOCKDEVICE_NAME_SCHEMA)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const DISK_LIST_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"A list of disk names, comma separated.")
|
||||||
|
.format(&ApiStringFormat::PropertyString(&DISK_ARRAY_SCHEMA))
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PASSWORD_SCHEMA: Schema = StringSchema::new("Password.")
|
||||||
|
.format(&PASSWORD_FORMAT)
|
||||||
|
.min_length(1)
|
||||||
|
.max_length(1024)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PBS_PASSWORD_SCHEMA: Schema = StringSchema::new("User Password.")
|
||||||
|
.format(&PASSWORD_FORMAT)
|
||||||
|
.min_length(5)
|
||||||
|
.max_length(64)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const REALM_ID_SCHEMA: Schema = StringSchema::new("Realm name.")
|
||||||
|
.format(&PROXMOX_SAFE_ID_FORMAT)
|
||||||
|
.min_length(2)
|
||||||
|
.max_length(32)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const FINGERPRINT_SHA256_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&FINGERPRINT_SHA256_REGEX);
|
||||||
|
|
||||||
|
pub const CERT_FINGERPRINT_SHA256_SCHEMA: Schema =
|
||||||
|
StringSchema::new("X509 certificate fingerprint (sha256).")
|
||||||
|
.format(&FINGERPRINT_SHA256_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PROXMOX_SAFE_ID_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&PROXMOX_SAFE_ID_REGEX);
|
||||||
|
|
||||||
|
pub const SINGLE_LINE_COMMENT_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&SINGLE_LINE_COMMENT_REGEX);
|
||||||
|
|
||||||
|
pub const SINGLE_LINE_COMMENT_SCHEMA: Schema = StringSchema::new("Comment (single line).")
|
||||||
|
.format(&SINGLE_LINE_COMMENT_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const SUBSCRIPTION_KEY_SCHEMA: Schema = StringSchema::new("Proxmox Backup Server subscription key.")
|
||||||
|
.format(&SUBSCRIPTION_KEY_FORMAT)
|
||||||
|
.min_length(15)
|
||||||
|
.max_length(16)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const SERVICE_ID_SCHEMA: Schema = StringSchema::new("Service ID.")
|
||||||
|
.max_length(256)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const PROXMOX_CONFIG_DIGEST_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Prevent changes if current configuration file has different \
|
||||||
|
SHA256 digest. This can be used to prevent concurrent \
|
||||||
|
modifications.",
|
||||||
|
)
|
||||||
|
.format(&PVE_CONFIG_DIGEST_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
/// API schema format definition for repository URLs
|
||||||
|
pub const BACKUP_REPO_URL: ApiStringFormat = ApiStringFormat::Pattern(&BACKUP_REPO_URL_REGEX);
|
||||||
|
|
||||||
|
|
||||||
|
// Complex type definitions
|
||||||
|
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Default, Serialize, Deserialize)]
|
||||||
|
/// Storage space usage information.
|
||||||
|
pub struct StorageStatus {
|
||||||
|
/// Total space (bytes).
|
||||||
|
pub total: u64,
|
||||||
|
/// Used space (bytes).
|
||||||
|
pub used: u64,
|
||||||
|
/// Available space (bytes).
|
||||||
|
pub avail: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const PASSWORD_HINT_SCHEMA: Schema = StringSchema::new("Password hint.")
|
||||||
|
.format(&SINGLE_LINE_COMMENT_FORMAT)
|
||||||
|
.min_length(1)
|
||||||
|
.max_length(64)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
|
||||||
|
#[api]
|
||||||
|
#[derive(Deserialize, Serialize)]
|
||||||
|
/// RSA public key information
|
||||||
|
pub struct RsaPubKeyInfo {
|
||||||
|
/// Path to key (if stored in a file)
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub path: Option<String>,
|
||||||
|
/// RSA exponent
|
||||||
|
pub exponent: String,
|
||||||
|
/// Hex-encoded RSA modulus
|
||||||
|
pub modulus: String,
|
||||||
|
/// Key (modulus) length in bits
|
||||||
|
pub length: usize,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl std::convert::TryFrom<openssl::rsa::Rsa<openssl::pkey::Public>> for RsaPubKeyInfo {
|
||||||
|
type Error = anyhow::Error;
|
||||||
|
|
||||||
|
fn try_from(value: openssl::rsa::Rsa<openssl::pkey::Public>) -> Result<Self, Self::Error> {
|
||||||
|
let modulus = value.n().to_hex_str()?.to_string();
|
||||||
|
let exponent = value.e().to_dec_str()?.to_string();
|
||||||
|
let length = value.size() as usize * 8;
|
||||||
|
|
||||||
|
Ok(Self {
|
||||||
|
path: None,
|
||||||
|
exponent,
|
||||||
|
modulus,
|
||||||
|
length,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "PascalCase")]
|
||||||
|
/// Describes a package for which an update is available.
|
||||||
|
pub struct APTUpdateInfo {
|
||||||
|
/// Package name
|
||||||
|
pub package: String,
|
||||||
|
/// Package title
|
||||||
|
pub title: String,
|
||||||
|
/// Package architecture
|
||||||
|
pub arch: String,
|
||||||
|
/// Human readable package description
|
||||||
|
pub description: String,
|
||||||
|
/// New version to be updated to
|
||||||
|
pub version: String,
|
||||||
|
/// Old version currently installed
|
||||||
|
pub old_version: String,
|
||||||
|
/// Package origin
|
||||||
|
pub origin: String,
|
||||||
|
/// Package priority in human-readable form
|
||||||
|
pub priority: String,
|
||||||
|
/// Package section
|
||||||
|
pub section: String,
|
||||||
|
/// URL under which the package's changelog can be retrieved
|
||||||
|
pub change_log_url: String,
|
||||||
|
/// Custom extra field for additional package information
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub extra_info: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
/// Node Power command type.
|
||||||
|
pub enum NodePowerCommand {
|
||||||
|
/// Restart the server
|
||||||
|
Reboot,
|
||||||
|
/// Shutdown the server
|
||||||
|
Shutdown,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Eq, PartialEq, Debug, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
pub enum TaskStateType {
|
||||||
|
/// Ok
|
||||||
|
OK,
|
||||||
|
/// Warning
|
||||||
|
Warning,
|
||||||
|
/// Error
|
||||||
|
Error,
|
||||||
|
/// Unknown
|
||||||
|
Unknown,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
upid: { schema: UPID::API_SCHEMA },
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize)]
|
||||||
|
/// Task properties.
|
||||||
|
pub struct TaskListItem {
|
||||||
|
pub upid: String,
|
||||||
|
/// The node name where the task is running on.
|
||||||
|
pub node: String,
|
||||||
|
/// The Unix PID
|
||||||
|
pub pid: i64,
|
||||||
|
/// The task start time (Epoch)
|
||||||
|
pub pstart: u64,
|
||||||
|
/// The task start time (Epoch)
|
||||||
|
pub starttime: i64,
|
||||||
|
/// Worker type (arbitrary ASCII string)
|
||||||
|
pub worker_type: String,
|
||||||
|
/// Worker ID (arbitrary ASCII string)
|
||||||
|
pub worker_id: Option<String>,
|
||||||
|
/// The authenticated entity who started the task
|
||||||
|
pub user: String,
|
||||||
|
/// The task end time (Epoch)
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub endtime: Option<i64>,
|
||||||
|
/// Task end status
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub status: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const NODE_TASKS_LIST_TASKS_RETURN_TYPE: ReturnType = ReturnType {
|
||||||
|
optional: false,
|
||||||
|
schema: &ArraySchema::new(
|
||||||
|
"A list of tasks.",
|
||||||
|
&TaskListItem::API_SCHEMA,
|
||||||
|
).schema(),
|
||||||
|
};
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Copy, Clone, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "UPPERCASE")]
|
||||||
|
/// RRD consolidation mode
|
||||||
|
pub enum RRDMode {
|
||||||
|
/// Maximum
|
||||||
|
Max,
|
||||||
|
/// Average
|
||||||
|
Average,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Copy, Clone, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
/// RRD time frame
|
||||||
|
pub enum RRDTimeFrame {
|
||||||
|
/// Hour
|
||||||
|
Hour,
|
||||||
|
/// Day
|
||||||
|
Day,
|
||||||
|
/// Week
|
||||||
|
Week,
|
||||||
|
/// Month
|
||||||
|
Month,
|
||||||
|
/// Year
|
||||||
|
Year,
|
||||||
|
/// Decade (10 years)
|
||||||
|
Decade,
|
||||||
|
}
|
308
pbs-api-types/src/network.rs
Normal file
308
pbs-api-types/src/network.rs
Normal file
@ -0,0 +1,308 @@
|
|||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::*;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
PROXMOX_SAFE_ID_REGEX,
|
||||||
|
IP_V4_FORMAT, IP_V6_FORMAT, IP_FORMAT,
|
||||||
|
CIDR_V4_FORMAT, CIDR_V6_FORMAT, CIDR_FORMAT,
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const NETWORK_INTERFACE_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&PROXMOX_SAFE_ID_REGEX);
|
||||||
|
|
||||||
|
pub const IP_V4_SCHEMA: Schema =
|
||||||
|
StringSchema::new("IPv4 address.")
|
||||||
|
.format(&IP_V4_FORMAT)
|
||||||
|
.max_length(15)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const IP_V6_SCHEMA: Schema =
|
||||||
|
StringSchema::new("IPv6 address.")
|
||||||
|
.format(&IP_V6_FORMAT)
|
||||||
|
.max_length(39)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const IP_SCHEMA: Schema =
|
||||||
|
StringSchema::new("IP (IPv4 or IPv6) address.")
|
||||||
|
.format(&IP_FORMAT)
|
||||||
|
.max_length(39)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const CIDR_V4_SCHEMA: Schema =
|
||||||
|
StringSchema::new("IPv4 address with netmask (CIDR notation).")
|
||||||
|
.format(&CIDR_V4_FORMAT)
|
||||||
|
.max_length(18)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const CIDR_V6_SCHEMA: Schema =
|
||||||
|
StringSchema::new("IPv6 address with netmask (CIDR notation).")
|
||||||
|
.format(&CIDR_V6_FORMAT)
|
||||||
|
.max_length(43)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const CIDR_SCHEMA: Schema =
|
||||||
|
StringSchema::new("IP address (IPv4 or IPv6) with netmask (CIDR notation).")
|
||||||
|
.format(&CIDR_FORMAT)
|
||||||
|
.max_length(43)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
/// Interface configuration method
|
||||||
|
pub enum NetworkConfigMethod {
|
||||||
|
/// Configuration is done manually using other tools
|
||||||
|
Manual,
|
||||||
|
/// Define interfaces with statically allocated addresses.
|
||||||
|
Static,
|
||||||
|
/// Obtain an address via DHCP
|
||||||
|
DHCP,
|
||||||
|
/// Define the loopback interface.
|
||||||
|
Loopback,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
#[allow(non_camel_case_types)]
|
||||||
|
#[repr(u8)]
|
||||||
|
/// Linux Bond Mode
|
||||||
|
pub enum LinuxBondMode {
|
||||||
|
/// Round-robin policy
|
||||||
|
balance_rr = 0,
|
||||||
|
/// Active-backup policy
|
||||||
|
active_backup = 1,
|
||||||
|
/// XOR policy
|
||||||
|
balance_xor = 2,
|
||||||
|
/// Broadcast policy
|
||||||
|
broadcast = 3,
|
||||||
|
/// IEEE 802.3ad Dynamic link aggregation
|
||||||
|
#[serde(rename = "802.3ad")]
|
||||||
|
ieee802_3ad = 4,
|
||||||
|
/// Adaptive transmit load balancing
|
||||||
|
balance_tlb = 5,
|
||||||
|
/// Adaptive load balancing
|
||||||
|
balance_alb = 6,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
#[allow(non_camel_case_types)]
|
||||||
|
#[repr(u8)]
|
||||||
|
/// Bond Transmit Hash Policy for LACP (802.3ad)
|
||||||
|
pub enum BondXmitHashPolicy {
|
||||||
|
/// Layer 2
|
||||||
|
layer2 = 0,
|
||||||
|
/// Layer 2+3
|
||||||
|
#[serde(rename = "layer2+3")]
|
||||||
|
layer2_3 = 1,
|
||||||
|
/// Layer 3+4
|
||||||
|
#[serde(rename = "layer3+4")]
|
||||||
|
layer3_4 = 2,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
/// Network interface type
|
||||||
|
pub enum NetworkInterfaceType {
|
||||||
|
/// Loopback
|
||||||
|
Loopback,
|
||||||
|
/// Physical Ethernet device
|
||||||
|
Eth,
|
||||||
|
/// Linux Bridge
|
||||||
|
Bridge,
|
||||||
|
/// Linux Bond
|
||||||
|
Bond,
|
||||||
|
/// Linux VLAN (eth.10)
|
||||||
|
Vlan,
|
||||||
|
/// Interface Alias (eth:1)
|
||||||
|
Alias,
|
||||||
|
/// Unknown interface type
|
||||||
|
Unknown,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const NETWORK_INTERFACE_NAME_SCHEMA: Schema = StringSchema::new("Network interface name.")
|
||||||
|
.format(&NETWORK_INTERFACE_FORMAT)
|
||||||
|
.min_length(1)
|
||||||
|
.max_length(libc::IFNAMSIZ-1)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const NETWORK_INTERFACE_ARRAY_SCHEMA: Schema = ArraySchema::new(
|
||||||
|
"Network interface list.", &NETWORK_INTERFACE_NAME_SCHEMA)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const NETWORK_INTERFACE_LIST_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"A list of network devices, comma separated.")
|
||||||
|
.format(&ApiStringFormat::PropertyString(&NETWORK_INTERFACE_ARRAY_SCHEMA))
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
name: {
|
||||||
|
schema: NETWORK_INTERFACE_NAME_SCHEMA,
|
||||||
|
},
|
||||||
|
"type": {
|
||||||
|
type: NetworkInterfaceType,
|
||||||
|
},
|
||||||
|
method: {
|
||||||
|
type: NetworkConfigMethod,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
method6: {
|
||||||
|
type: NetworkConfigMethod,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
cidr: {
|
||||||
|
schema: CIDR_V4_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
cidr6: {
|
||||||
|
schema: CIDR_V6_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
gateway: {
|
||||||
|
schema: IP_V4_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
gateway6: {
|
||||||
|
schema: IP_V6_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
options: {
|
||||||
|
description: "Option list (inet)",
|
||||||
|
type: Array,
|
||||||
|
items: {
|
||||||
|
description: "Optional attribute line.",
|
||||||
|
type: String,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
options6: {
|
||||||
|
description: "Option list (inet6)",
|
||||||
|
type: Array,
|
||||||
|
items: {
|
||||||
|
description: "Optional attribute line.",
|
||||||
|
type: String,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
comments: {
|
||||||
|
description: "Comments (inet, may span multiple lines)",
|
||||||
|
type: String,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
comments6: {
|
||||||
|
description: "Comments (inet6, may span multiple lines)",
|
||||||
|
type: String,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
bridge_ports: {
|
||||||
|
schema: NETWORK_INTERFACE_ARRAY_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
slaves: {
|
||||||
|
schema: NETWORK_INTERFACE_ARRAY_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
bond_mode: {
|
||||||
|
type: LinuxBondMode,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"bond-primary": {
|
||||||
|
schema: NETWORK_INTERFACE_NAME_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
bond_xmit_hash_policy: {
|
||||||
|
type: BondXmitHashPolicy,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
|
/// Network Interface configuration
|
||||||
|
pub struct Interface {
|
||||||
|
/// Autostart interface
|
||||||
|
#[serde(rename = "autostart")]
|
||||||
|
pub autostart: bool,
|
||||||
|
/// Interface is active (UP)
|
||||||
|
pub active: bool,
|
||||||
|
/// Interface name
|
||||||
|
pub name: String,
|
||||||
|
/// Interface type
|
||||||
|
#[serde(rename = "type")]
|
||||||
|
pub interface_type: NetworkInterfaceType,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub method: Option<NetworkConfigMethod>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub method6: Option<NetworkConfigMethod>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
/// IPv4 address with netmask
|
||||||
|
pub cidr: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
/// IPv4 gateway
|
||||||
|
pub gateway: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
/// IPv6 address with netmask
|
||||||
|
pub cidr6: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
/// IPv6 gateway
|
||||||
|
pub gateway6: Option<String>,
|
||||||
|
|
||||||
|
#[serde(skip_serializing_if="Vec::is_empty")]
|
||||||
|
pub options: Vec<String>,
|
||||||
|
#[serde(skip_serializing_if="Vec::is_empty")]
|
||||||
|
pub options6: Vec<String>,
|
||||||
|
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comments: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comments6: Option<String>,
|
||||||
|
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
/// Maximum Transmission Unit
|
||||||
|
pub mtu: Option<u64>,
|
||||||
|
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub bridge_ports: Option<Vec<String>>,
|
||||||
|
/// Enable bridge vlan support.
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub bridge_vlan_aware: Option<bool>,
|
||||||
|
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub slaves: Option<Vec<String>>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub bond_mode: Option<LinuxBondMode>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
#[serde(rename = "bond-primary")]
|
||||||
|
pub bond_primary: Option<String>,
|
||||||
|
pub bond_xmit_hash_policy: Option<BondXmitHashPolicy>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Interface {
|
||||||
|
pub fn new(name: String) -> Self {
|
||||||
|
Self {
|
||||||
|
name,
|
||||||
|
interface_type: NetworkInterfaceType::Unknown,
|
||||||
|
autostart: false,
|
||||||
|
active: false,
|
||||||
|
method: None,
|
||||||
|
method6: None,
|
||||||
|
cidr: None,
|
||||||
|
gateway: None,
|
||||||
|
cidr6: None,
|
||||||
|
gateway6: None,
|
||||||
|
options: Vec::new(),
|
||||||
|
options6: Vec::new(),
|
||||||
|
comments: None,
|
||||||
|
comments6: None,
|
||||||
|
mtu: None,
|
||||||
|
bridge_ports: None,
|
||||||
|
bridge_vlan_aware: None,
|
||||||
|
slaves: None,
|
||||||
|
bond_mode: None,
|
||||||
|
bond_primary: None,
|
||||||
|
bond_xmit_hash_policy: None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
121
pbs-api-types/src/openid.rs
Normal file
121
pbs-api-types/src/openid.rs
Normal file
@ -0,0 +1,121 @@
|
|||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::{
|
||||||
|
api, ApiStringFormat, ArraySchema, Schema, StringSchema, Updater,
|
||||||
|
};
|
||||||
|
|
||||||
|
use super::{
|
||||||
|
PROXMOX_SAFE_ID_REGEX, PROXMOX_SAFE_ID_FORMAT, REALM_ID_SCHEMA,
|
||||||
|
SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const OPENID_SCOPE_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&PROXMOX_SAFE_ID_REGEX);
|
||||||
|
|
||||||
|
pub const OPENID_SCOPE_SCHEMA: Schema = StringSchema::new("OpenID Scope Name.")
|
||||||
|
.format(&OPENID_SCOPE_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const OPENID_SCOPE_ARRAY_SCHEMA: Schema = ArraySchema::new(
|
||||||
|
"Array of OpenId Scopes.", &OPENID_SCOPE_SCHEMA).schema();
|
||||||
|
|
||||||
|
pub const OPENID_SCOPE_LIST_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::PropertyString(&OPENID_SCOPE_ARRAY_SCHEMA);
|
||||||
|
|
||||||
|
pub const OPENID_DEFAILT_SCOPE_LIST: &'static str = "email profile";
|
||||||
|
pub const OPENID_SCOPE_LIST_SCHEMA: Schema = StringSchema::new("OpenID Scope List")
|
||||||
|
.format(&OPENID_SCOPE_LIST_FORMAT)
|
||||||
|
.default(OPENID_DEFAILT_SCOPE_LIST)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const OPENID_ACR_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&PROXMOX_SAFE_ID_REGEX);
|
||||||
|
|
||||||
|
pub const OPENID_ACR_SCHEMA: Schema = StringSchema::new("OpenID Authentication Context Class Reference.")
|
||||||
|
.format(&OPENID_SCOPE_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const OPENID_ACR_ARRAY_SCHEMA: Schema = ArraySchema::new(
|
||||||
|
"Array of OpenId ACRs.", &OPENID_ACR_SCHEMA).schema();
|
||||||
|
|
||||||
|
pub const OPENID_ACR_LIST_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::PropertyString(&OPENID_ACR_ARRAY_SCHEMA);
|
||||||
|
|
||||||
|
pub const OPENID_ACR_LIST_SCHEMA: Schema = StringSchema::new("OpenID ACR List")
|
||||||
|
.format(&OPENID_ACR_LIST_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const OPENID_USERNAME_CLAIM_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Use the value of this attribute/claim as unique user name. It \
|
||||||
|
is up to the identity provider to guarantee the uniqueness. The \
|
||||||
|
OpenID specification only guarantees that Subject ('sub') is \
|
||||||
|
unique. Also make sure that the user is not allowed to change that \
|
||||||
|
attribute by himself!")
|
||||||
|
.max_length(64)
|
||||||
|
.min_length(1)
|
||||||
|
.format(&PROXMOX_SAFE_ID_FORMAT) .schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
realm: {
|
||||||
|
schema: REALM_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
"client-key": {
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"scopes": {
|
||||||
|
schema: OPENID_SCOPE_LIST_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"acr-values": {
|
||||||
|
schema: OPENID_ACR_LIST_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
prompt: {
|
||||||
|
description: "OpenID Prompt",
|
||||||
|
type: String,
|
||||||
|
format: &PROXMOX_SAFE_ID_FORMAT,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
autocreate: {
|
||||||
|
optional: true,
|
||||||
|
default: false,
|
||||||
|
},
|
||||||
|
"username-claim": {
|
||||||
|
schema: OPENID_USERNAME_CLAIM_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize, Deserialize, Updater)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// OpenID configuration properties.
|
||||||
|
pub struct OpenIdRealmConfig {
|
||||||
|
#[updater(skip)]
|
||||||
|
pub realm: String,
|
||||||
|
/// OpenID Issuer Url
|
||||||
|
pub issuer_url: String,
|
||||||
|
/// OpenID Client ID
|
||||||
|
pub client_id: String,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub scopes: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub acr_values: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub prompt: Option<String>,
|
||||||
|
/// OpenID Client Key
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub client_key: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
/// Automatically create users if they do not exist.
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub autocreate: Option<bool>,
|
||||||
|
#[updater(skip)]
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub username_claim: Option<String>,
|
||||||
|
}
|
86
pbs-api-types/src/remote.rs
Normal file
86
pbs-api-types/src/remote.rs
Normal file
@ -0,0 +1,86 @@
|
|||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use super::*;
|
||||||
|
use proxmox_schema::*;
|
||||||
|
|
||||||
|
pub const REMOTE_PASSWORD_SCHEMA: Schema = StringSchema::new("Password or auth token for remote host.")
|
||||||
|
.format(&PASSWORD_FORMAT)
|
||||||
|
.min_length(1)
|
||||||
|
.max_length(1024)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const REMOTE_PASSWORD_BASE64_SCHEMA: Schema = StringSchema::new("Password or auth token for remote host (stored as base64 string).")
|
||||||
|
.format(&PASSWORD_FORMAT)
|
||||||
|
.min_length(1)
|
||||||
|
.max_length(1024)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const REMOTE_ID_SCHEMA: Schema = StringSchema::new("Remote ID.")
|
||||||
|
.format(&PROXMOX_SAFE_ID_FORMAT)
|
||||||
|
.min_length(3)
|
||||||
|
.max_length(32)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
host: {
|
||||||
|
schema: DNS_NAME_OR_IP_SCHEMA,
|
||||||
|
},
|
||||||
|
port: {
|
||||||
|
optional: true,
|
||||||
|
description: "The (optional) port",
|
||||||
|
type: u16,
|
||||||
|
},
|
||||||
|
"auth-id": {
|
||||||
|
type: Authid,
|
||||||
|
},
|
||||||
|
fingerprint: {
|
||||||
|
optional: true,
|
||||||
|
schema: CERT_FINGERPRINT_SHA256_SCHEMA,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize,Updater)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Remote configuration properties.
|
||||||
|
pub struct RemoteConfig {
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
pub host: String,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub port: Option<u16>,
|
||||||
|
pub auth_id: Authid,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub fingerprint: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
name: {
|
||||||
|
schema: REMOTE_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
config: {
|
||||||
|
type: RemoteConfig,
|
||||||
|
},
|
||||||
|
password: {
|
||||||
|
schema: REMOTE_PASSWORD_BASE64_SCHEMA,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Remote properties.
|
||||||
|
pub struct Remote {
|
||||||
|
pub name: String,
|
||||||
|
// Note: The stored password is base64 encoded
|
||||||
|
#[serde(skip_serializing_if="String::is_empty")]
|
||||||
|
#[serde(with = "proxmox::tools::serde::string_as_base64")]
|
||||||
|
pub password: String,
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub config: RemoteConfig,
|
||||||
|
}
|
@ -2,21 +2,11 @@
|
|||||||
|
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use proxmox::api::{
|
use proxmox_schema::{
|
||||||
api,
|
api, ApiStringFormat, ArraySchema, IntegerSchema, Schema, StringSchema, Updater,
|
||||||
schema::{
|
|
||||||
Schema,
|
|
||||||
ApiStringFormat,
|
|
||||||
ArraySchema,
|
|
||||||
IntegerSchema,
|
|
||||||
StringSchema,
|
|
||||||
},
|
|
||||||
};
|
};
|
||||||
|
|
||||||
use crate::api2::types::{
|
use crate::{OptionalDeviceIdentification, PROXMOX_SAFE_ID_FORMAT};
|
||||||
PROXMOX_SAFE_ID_FORMAT,
|
|
||||||
OptionalDeviceIdentification,
|
|
||||||
};
|
|
||||||
|
|
||||||
pub const CHANGER_NAME_SCHEMA: Schema = StringSchema::new("Tape Changer Identifier.")
|
pub const CHANGER_NAME_SCHEMA: Schema = StringSchema::new("Tape Changer Identifier.")
|
||||||
.format(&PROXMOX_SAFE_ID_FORMAT)
|
.format(&PROXMOX_SAFE_ID_FORMAT)
|
||||||
@ -24,9 +14,8 @@ pub const CHANGER_NAME_SCHEMA: Schema = StringSchema::new("Tape Changer Identifi
|
|||||||
.max_length(32)
|
.max_length(32)
|
||||||
.schema();
|
.schema();
|
||||||
|
|
||||||
pub const SCSI_CHANGER_PATH_SCHEMA: Schema = StringSchema::new(
|
pub const SCSI_CHANGER_PATH_SCHEMA: Schema =
|
||||||
"Path to Linux generic SCSI device (e.g. '/dev/sg4')")
|
StringSchema::new("Path to Linux generic SCSI device (e.g. '/dev/sg4')").schema();
|
||||||
.schema();
|
|
||||||
|
|
||||||
pub const MEDIA_LABEL_SCHEMA: Schema = StringSchema::new("Media Label/Barcode.")
|
pub const MEDIA_LABEL_SCHEMA: Schema = StringSchema::new("Media Label/Barcode.")
|
||||||
.format(&PROXMOX_SAFE_ID_FORMAT)
|
.format(&PROXMOX_SAFE_ID_FORMAT)
|
||||||
@ -35,16 +24,18 @@ pub const MEDIA_LABEL_SCHEMA: Schema = StringSchema::new("Media Label/Barcode.")
|
|||||||
.schema();
|
.schema();
|
||||||
|
|
||||||
pub const SLOT_ARRAY_SCHEMA: Schema = ArraySchema::new(
|
pub const SLOT_ARRAY_SCHEMA: Schema = ArraySchema::new(
|
||||||
"Slot list.", &IntegerSchema::new("Slot number")
|
"Slot list.",
|
||||||
.minimum(1)
|
&IntegerSchema::new("Slot number").minimum(1).schema(),
|
||||||
.schema())
|
)
|
||||||
.schema();
|
.schema();
|
||||||
|
|
||||||
pub const EXPORT_SLOT_LIST_SCHEMA: Schema = StringSchema::new("\
|
pub const EXPORT_SLOT_LIST_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"\
|
||||||
A list of slot numbers, comma separated. Those slots are reserved for
|
A list of slot numbers, comma separated. Those slots are reserved for
|
||||||
Import/Export, i.e. any media in those slots are considered to be
|
Import/Export, i.e. any media in those slots are considered to be
|
||||||
'offline'.
|
'offline'.
|
||||||
")
|
",
|
||||||
|
)
|
||||||
.format(&ApiStringFormat::PropertyString(&SLOT_ARRAY_SCHEMA))
|
.format(&ApiStringFormat::PropertyString(&SLOT_ARRAY_SCHEMA))
|
||||||
.schema();
|
.schema();
|
||||||
|
|
||||||
@ -62,13 +53,14 @@ Import/Export, i.e. any media in those slots are considered to be
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
)]
|
)]
|
||||||
#[derive(Serialize,Deserialize)]
|
#[derive(Serialize, Deserialize, Updater)]
|
||||||
#[serde(rename_all = "kebab-case")]
|
#[serde(rename_all = "kebab-case")]
|
||||||
/// SCSI tape changer
|
/// SCSI tape changer
|
||||||
pub struct ScsiTapeChanger {
|
pub struct ScsiTapeChanger {
|
||||||
|
#[updater(skip)]
|
||||||
pub name: String,
|
pub name: String,
|
||||||
pub path: String,
|
pub path: String,
|
||||||
#[serde(skip_serializing_if="Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub export_slots: Option<String>,
|
pub export_slots: Option<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -82,7 +74,7 @@ pub struct ScsiTapeChanger {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
)]
|
)]
|
||||||
#[derive(Serialize,Deserialize)]
|
#[derive(Serialize, Deserialize)]
|
||||||
#[serde(rename_all = "kebab-case")]
|
#[serde(rename_all = "kebab-case")]
|
||||||
/// Changer config with optional device identification attributes
|
/// Changer config with optional device identification attributes
|
||||||
pub struct ChangerListEntry {
|
pub struct ChangerListEntry {
|
||||||
@ -93,7 +85,7 @@ pub struct ChangerListEntry {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[api()]
|
#[api()]
|
||||||
#[derive(Serialize,Deserialize)]
|
#[derive(Serialize, Deserialize)]
|
||||||
#[serde(rename_all = "kebab-case")]
|
#[serde(rename_all = "kebab-case")]
|
||||||
/// Mtx Entry Kind
|
/// Mtx Entry Kind
|
||||||
pub enum MtxEntryKind {
|
pub enum MtxEntryKind {
|
||||||
@ -116,7 +108,7 @@ pub enum MtxEntryKind {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
)]
|
)]
|
||||||
#[derive(Serialize,Deserialize)]
|
#[derive(Serialize, Deserialize)]
|
||||||
#[serde(rename_all = "kebab-case")]
|
#[serde(rename_all = "kebab-case")]
|
||||||
/// Mtx Status Entry
|
/// Mtx Status Entry
|
||||||
pub struct MtxStatusEntry {
|
pub struct MtxStatusEntry {
|
||||||
@ -124,12 +116,12 @@ pub struct MtxStatusEntry {
|
|||||||
/// The ID of the slot or drive
|
/// The ID of the slot or drive
|
||||||
pub entry_id: u64,
|
pub entry_id: u64,
|
||||||
/// The media label (volume tag) if the slot/drive is full
|
/// The media label (volume tag) if the slot/drive is full
|
||||||
#[serde(skip_serializing_if="Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub label_text: Option<String>,
|
pub label_text: Option<String>,
|
||||||
/// The slot the drive was loaded from
|
/// The slot the drive was loaded from
|
||||||
#[serde(skip_serializing_if="Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub loaded_slot: Option<u64>,
|
pub loaded_slot: Option<u64>,
|
||||||
/// The current state of the drive
|
/// The current state of the drive
|
||||||
#[serde(skip_serializing_if="Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub state: Option<String>,
|
pub state: Option<String>,
|
||||||
}
|
}
|
@ -1,6 +1,6 @@
|
|||||||
use ::serde::{Deserialize, Serialize};
|
use ::serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use proxmox::api::api;
|
use proxmox_schema::api;
|
||||||
|
|
||||||
#[api()]
|
#[api()]
|
||||||
#[derive(Serialize,Deserialize)]
|
#[derive(Serialize,Deserialize)]
|
@ -4,12 +4,9 @@ use std::convert::TryFrom;
|
|||||||
use anyhow::{bail, Error};
|
use anyhow::{bail, Error};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use proxmox::api::{
|
use proxmox_schema::{api, Schema, IntegerSchema, StringSchema, Updater};
|
||||||
api,
|
|
||||||
schema::{Schema, IntegerSchema, StringSchema},
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::api2::types::{
|
use crate::{
|
||||||
PROXMOX_SAFE_ID_FORMAT,
|
PROXMOX_SAFE_ID_FORMAT,
|
||||||
CHANGER_NAME_SCHEMA,
|
CHANGER_NAME_SCHEMA,
|
||||||
OptionalDeviceIdentification,
|
OptionalDeviceIdentification,
|
||||||
@ -28,7 +25,7 @@ pub const LTO_DRIVE_PATH_SCHEMA: Schema = StringSchema::new(
|
|||||||
pub const CHANGER_DRIVENUM_SCHEMA: Schema = IntegerSchema::new(
|
pub const CHANGER_DRIVENUM_SCHEMA: Schema = IntegerSchema::new(
|
||||||
"Associated changer drive number (requires option changer)")
|
"Associated changer drive number (requires option changer)")
|
||||||
.minimum(0)
|
.minimum(0)
|
||||||
.maximum(8)
|
.maximum(255)
|
||||||
.default(0)
|
.default(0)
|
||||||
.schema();
|
.schema();
|
||||||
|
|
||||||
@ -69,10 +66,11 @@ pub struct VirtualTapeDrive {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
)]
|
)]
|
||||||
#[derive(Serialize,Deserialize)]
|
#[derive(Serialize,Deserialize,Updater)]
|
||||||
#[serde(rename_all = "kebab-case")]
|
#[serde(rename_all = "kebab-case")]
|
||||||
/// Lto SCSI tape driver
|
/// Lto SCSI tape driver
|
||||||
pub struct LtoTapeDrive {
|
pub struct LtoTapeDrive {
|
||||||
|
#[updater(skip)]
|
||||||
pub name: String,
|
pub name: String,
|
||||||
pub path: String,
|
pub path: String,
|
||||||
#[serde(skip_serializing_if="Option::is_none")]
|
#[serde(skip_serializing_if="Option::is_none")]
|
@ -1,17 +1,44 @@
|
|||||||
use ::serde::{Deserialize, Serialize};
|
use ::serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use proxmox::{
|
use proxmox_schema::*;
|
||||||
api::api,
|
use proxmox_uuid::Uuid;
|
||||||
tools::Uuid,
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::api2::types::{
|
use crate::{
|
||||||
MEDIA_UUID_SCHEMA,
|
UUID_FORMAT,
|
||||||
MEDIA_SET_UUID_SCHEMA,
|
|
||||||
MediaStatus,
|
MediaStatus,
|
||||||
MediaLocation,
|
MediaLocation,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
pub const MEDIA_SET_UUID_SCHEMA: Schema =
|
||||||
|
StringSchema::new("MediaSet Uuid (We use the all-zero Uuid to reseve an empty media for a specific pool).")
|
||||||
|
.format(&UUID_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const MEDIA_UUID_SCHEMA: Schema =
|
||||||
|
StringSchema::new("Media Uuid.")
|
||||||
|
.format(&UUID_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"media-set-uuid": {
|
||||||
|
schema: MEDIA_SET_UUID_SCHEMA,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Media Set list entry
|
||||||
|
pub struct MediaSetListEntry {
|
||||||
|
/// Media set name
|
||||||
|
pub media_set_name: String,
|
||||||
|
pub media_set_uuid: Uuid,
|
||||||
|
/// MediaSet creation time stamp
|
||||||
|
pub media_set_ctime: i64,
|
||||||
|
/// Media Pool
|
||||||
|
pub pool: String,
|
||||||
|
}
|
||||||
|
|
||||||
#[api(
|
#[api(
|
||||||
properties: {
|
properties: {
|
||||||
location: {
|
location: {
|
@ -1,18 +1,8 @@
|
|||||||
use anyhow::{bail, Error};
|
use anyhow::{bail, Error};
|
||||||
|
|
||||||
use proxmox::api::{
|
use proxmox_schema::{parse_simple_value, ApiStringFormat, Schema, StringSchema};
|
||||||
schema::{
|
|
||||||
Schema,
|
|
||||||
StringSchema,
|
|
||||||
ApiStringFormat,
|
|
||||||
parse_simple_value,
|
|
||||||
},
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::api2::types::{
|
use crate::{CHANGER_NAME_SCHEMA, PROXMOX_SAFE_ID_FORMAT};
|
||||||
PROXMOX_SAFE_ID_FORMAT,
|
|
||||||
CHANGER_NAME_SCHEMA,
|
|
||||||
};
|
|
||||||
|
|
||||||
pub const VAULT_NAME_SCHEMA: Schema = StringSchema::new("Vault name.")
|
pub const VAULT_NAME_SCHEMA: Schema = StringSchema::new("Vault name.")
|
||||||
.format(&PROXMOX_SAFE_ID_FORMAT)
|
.format(&PROXMOX_SAFE_ID_FORMAT)
|
||||||
@ -35,28 +25,27 @@ pub enum MediaLocation {
|
|||||||
proxmox::forward_deserialize_to_from_str!(MediaLocation);
|
proxmox::forward_deserialize_to_from_str!(MediaLocation);
|
||||||
proxmox::forward_serialize_to_display!(MediaLocation);
|
proxmox::forward_serialize_to_display!(MediaLocation);
|
||||||
|
|
||||||
impl MediaLocation {
|
impl proxmox_schema::ApiType for MediaLocation {
|
||||||
pub const API_SCHEMA: Schema = StringSchema::new(
|
const API_SCHEMA: Schema = StringSchema::new(
|
||||||
"Media location (e.g. 'offline', 'online-<changer_name>', 'vault-<vault_name>')")
|
"Media location (e.g. 'offline', 'online-<changer_name>', 'vault-<vault_name>')",
|
||||||
.format(&ApiStringFormat::VerifyFn(|text| {
|
)
|
||||||
let location: MediaLocation = text.parse()?;
|
.format(&ApiStringFormat::VerifyFn(|text| {
|
||||||
match location {
|
let location: MediaLocation = text.parse()?;
|
||||||
MediaLocation::Online(ref changer) => {
|
match location {
|
||||||
parse_simple_value(changer, &CHANGER_NAME_SCHEMA)?;
|
MediaLocation::Online(ref changer) => {
|
||||||
}
|
parse_simple_value(changer, &CHANGER_NAME_SCHEMA)?;
|
||||||
MediaLocation::Vault(ref vault) => {
|
|
||||||
parse_simple_value(vault, &VAULT_NAME_SCHEMA)?;
|
|
||||||
}
|
|
||||||
MediaLocation::Offline => { /* OK */}
|
|
||||||
}
|
}
|
||||||
Ok(())
|
MediaLocation::Vault(ref vault) => {
|
||||||
}))
|
parse_simple_value(vault, &VAULT_NAME_SCHEMA)?;
|
||||||
.schema();
|
}
|
||||||
|
MediaLocation::Offline => { /* OK */ }
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}))
|
||||||
|
.schema();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
impl std::fmt::Display for MediaLocation {
|
impl std::fmt::Display for MediaLocation {
|
||||||
|
|
||||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
match self {
|
match self {
|
||||||
MediaLocation::Offline => {
|
MediaLocation::Offline => {
|
@ -4,28 +4,20 @@
|
|||||||
//! so we cannot use them directly for the API. Instead, we represent
|
//! so we cannot use them directly for the API. Instead, we represent
|
||||||
//! them as String.
|
//! them as String.
|
||||||
|
|
||||||
use anyhow::Error;
|
|
||||||
use std::str::FromStr;
|
use std::str::FromStr;
|
||||||
|
|
||||||
|
use anyhow::Error;
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use proxmox::api::{
|
use proxmox_schema::{api, Schema, StringSchema, ApiStringFormat, Updater};
|
||||||
api,
|
|
||||||
schema::{Schema, StringSchema, ApiStringFormat},
|
use proxmox_time::{parse_calendar_event, parse_time_span, CalendarEvent, TimeSpan};
|
||||||
};
|
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
tools::systemd::time::{
|
PROXMOX_SAFE_ID_FORMAT,
|
||||||
CalendarEvent,
|
SINGLE_LINE_COMMENT_FORMAT,
|
||||||
TimeSpan,
|
SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
parse_time_span,
|
TAPE_ENCRYPTION_KEY_FINGERPRINT_SCHEMA,
|
||||||
parse_calendar_event,
|
|
||||||
},
|
|
||||||
api2::types::{
|
|
||||||
PROXMOX_SAFE_ID_FORMAT,
|
|
||||||
SINGLE_LINE_COMMENT_FORMAT,
|
|
||||||
SINGLE_LINE_COMMENT_SCHEMA,
|
|
||||||
TAPE_ENCRYPTION_KEY_FINGERPRINT_SCHEMA,
|
|
||||||
},
|
|
||||||
};
|
};
|
||||||
|
|
||||||
pub const MEDIA_POOL_NAME_SCHEMA: Schema = StringSchema::new("Media pool name.")
|
pub const MEDIA_POOL_NAME_SCHEMA: Schema = StringSchema::new("Media pool name.")
|
||||||
@ -138,10 +130,11 @@ impl std::str::FromStr for RetentionPolicy {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
)]
|
)]
|
||||||
#[derive(Serialize,Deserialize)]
|
#[derive(Serialize,Deserialize,Updater)]
|
||||||
/// Media pool configuration
|
/// Media pool configuration
|
||||||
pub struct MediaPoolConfig {
|
pub struct MediaPoolConfig {
|
||||||
/// The pool name
|
/// The pool name
|
||||||
|
#[updater(skip)]
|
||||||
pub name: String,
|
pub name: String,
|
||||||
/// Media Set allocation policy
|
/// Media Set allocation policy
|
||||||
#[serde(skip_serializing_if="Option::is_none")]
|
#[serde(skip_serializing_if="Option::is_none")]
|
@ -1,6 +1,6 @@
|
|||||||
use ::serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use proxmox::api::api;
|
use proxmox_schema::api;
|
||||||
|
|
||||||
#[api()]
|
#[api()]
|
||||||
/// Media status
|
/// Media status
|
91
pbs-api-types/src/tape/mod.rs
Normal file
91
pbs-api-types/src/tape/mod.rs
Normal file
@ -0,0 +1,91 @@
|
|||||||
|
//! Types for tape backup API
|
||||||
|
|
||||||
|
mod device;
|
||||||
|
pub use device::*;
|
||||||
|
|
||||||
|
mod changer;
|
||||||
|
pub use changer::*;
|
||||||
|
|
||||||
|
mod drive;
|
||||||
|
pub use drive::*;
|
||||||
|
|
||||||
|
mod media_pool;
|
||||||
|
pub use media_pool::*;
|
||||||
|
|
||||||
|
mod media_status;
|
||||||
|
pub use media_status::*;
|
||||||
|
|
||||||
|
mod media_location;
|
||||||
|
|
||||||
|
pub use media_location::*;
|
||||||
|
|
||||||
|
mod media;
|
||||||
|
pub use media::*;
|
||||||
|
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::{api, const_regex, Schema, StringSchema, ApiStringFormat};
|
||||||
|
use proxmox_uuid::Uuid;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
FINGERPRINT_SHA256_FORMAT, BACKUP_ID_SCHEMA, BACKUP_TYPE_SCHEMA,
|
||||||
|
};
|
||||||
|
|
||||||
|
const_regex!{
|
||||||
|
pub TAPE_RESTORE_SNAPSHOT_REGEX = concat!(r"^", PROXMOX_SAFE_ID_REGEX_STR!(), r":", SNAPSHOT_PATH_REGEX_STR!(), r"$");
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const TAPE_RESTORE_SNAPSHOT_FORMAT: ApiStringFormat =
|
||||||
|
ApiStringFormat::Pattern(&TAPE_RESTORE_SNAPSHOT_REGEX);
|
||||||
|
|
||||||
|
pub const TAPE_ENCRYPTION_KEY_FINGERPRINT_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Tape encryption key fingerprint (sha256)."
|
||||||
|
)
|
||||||
|
.format(&FINGERPRINT_SHA256_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const TAPE_RESTORE_SNAPSHOT_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"A snapshot in the format: 'store:type/id/time")
|
||||||
|
.format(&TAPE_RESTORE_SNAPSHOT_FORMAT)
|
||||||
|
.type_text("store:type/id/time")
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
pool: {
|
||||||
|
schema: MEDIA_POOL_NAME_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"label-text": {
|
||||||
|
schema: MEDIA_LABEL_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"media": {
|
||||||
|
schema: MEDIA_UUID_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"media-set": {
|
||||||
|
schema: MEDIA_SET_UUID_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"backup-type": {
|
||||||
|
schema: BACKUP_TYPE_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"backup-id": {
|
||||||
|
schema: BACKUP_ID_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// Content list filter parameters
|
||||||
|
pub struct MediaContentListFilter {
|
||||||
|
pub pool: Option<String>,
|
||||||
|
pub label_text: Option<String>,
|
||||||
|
pub media: Option<Uuid>,
|
||||||
|
pub media_set: Option<Uuid>,
|
||||||
|
pub backup_type: Option<String>,
|
||||||
|
pub backup_id: Option<String>,
|
||||||
|
}
|
122
pbs-api-types/src/traffic_control.rs
Normal file
122
pbs-api-types/src/traffic_control.rs
Normal file
@ -0,0 +1,122 @@
|
|||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::{api, Schema, IntegerSchema, StringSchema, Updater};
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
HumanByte, CIDR_SCHEMA, DAILY_DURATION_FORMAT,
|
||||||
|
PROXMOX_SAFE_ID_FORMAT, SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const TRAFFIC_CONTROL_TIMEFRAME_SCHEMA: Schema = StringSchema::new(
|
||||||
|
"Timeframe to specify when the rule is actice.")
|
||||||
|
.format(&DAILY_DURATION_FORMAT)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const TRAFFIC_CONTROL_ID_SCHEMA: Schema = StringSchema::new("Rule ID.")
|
||||||
|
.format(&PROXMOX_SAFE_ID_FORMAT)
|
||||||
|
.min_length(3)
|
||||||
|
.max_length(32)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const TRAFFIC_CONTROL_RATE_SCHEMA: Schema = IntegerSchema::new(
|
||||||
|
"Rate limit (for Token bucket filter) in bytes/second.")
|
||||||
|
.minimum(100_000)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const TRAFFIC_CONTROL_BURST_SCHEMA: Schema = IntegerSchema::new(
|
||||||
|
"Size of the token bucket (for Token bucket filter) in bytes.")
|
||||||
|
.minimum(1000)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
"rate-in": {
|
||||||
|
type: HumanByte,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"burst-in": {
|
||||||
|
type: HumanByte,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"rate-out": {
|
||||||
|
type: HumanByte,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
"burst-out": {
|
||||||
|
type: HumanByte,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize,Default,Clone,Updater)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Rate Limit Configuration
|
||||||
|
pub struct RateLimitConfig {
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub rate_in: Option<HumanByte>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub burst_in: Option<HumanByte>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub rate_out: Option<HumanByte>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub burst_out: Option<HumanByte>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl RateLimitConfig {
|
||||||
|
pub fn with_same_inout(rate: Option<HumanByte>, burst: Option<HumanByte>) -> Self {
|
||||||
|
Self {
|
||||||
|
rate_in: rate,
|
||||||
|
burst_in: burst,
|
||||||
|
rate_out: rate,
|
||||||
|
burst_out: burst,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
name: {
|
||||||
|
schema: TRAFFIC_CONTROL_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
limit: {
|
||||||
|
type: RateLimitConfig,
|
||||||
|
},
|
||||||
|
network: {
|
||||||
|
type: Array,
|
||||||
|
items: {
|
||||||
|
schema: CIDR_SCHEMA,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
timeframe: {
|
||||||
|
type: Array,
|
||||||
|
items: {
|
||||||
|
schema: TRAFFIC_CONTROL_TIMEFRAME_SCHEMA,
|
||||||
|
},
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize, Updater)]
|
||||||
|
#[serde(rename_all = "kebab-case")]
|
||||||
|
/// Traffic control rule
|
||||||
|
pub struct TrafficControlRule {
|
||||||
|
#[updater(skip)]
|
||||||
|
pub name: String,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
/// Rule applies to Source IPs within this networks
|
||||||
|
pub network: Vec<String>,
|
||||||
|
#[serde(flatten)]
|
||||||
|
pub limit: RateLimitConfig,
|
||||||
|
// fixme: expose this?
|
||||||
|
// /// Bandwidth is shared accross all connections
|
||||||
|
// #[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
// pub shared: Option<bool>,
|
||||||
|
/// Enable the rule at specific times
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub timeframe: Option<Vec<String>>,
|
||||||
|
}
|
207
pbs-api-types/src/user.rs
Normal file
207
pbs-api-types/src/user.rs
Normal file
@ -0,0 +1,207 @@
|
|||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::{
|
||||||
|
api, BooleanSchema, IntegerSchema, Schema, StringSchema, Updater,
|
||||||
|
};
|
||||||
|
|
||||||
|
use super::{SINGLE_LINE_COMMENT_FORMAT, SINGLE_LINE_COMMENT_SCHEMA};
|
||||||
|
use super::userid::{Authid, Userid, PROXMOX_TOKEN_ID_SCHEMA};
|
||||||
|
|
||||||
|
pub const ENABLE_USER_SCHEMA: Schema = BooleanSchema::new(
|
||||||
|
"Enable the account (default). You can set this to '0' to disable the account.")
|
||||||
|
.default(true)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const EXPIRE_USER_SCHEMA: Schema = IntegerSchema::new(
|
||||||
|
"Account expiration date (seconds since epoch). '0' means no expiration date.")
|
||||||
|
.default(0)
|
||||||
|
.minimum(0)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const FIRST_NAME_SCHEMA: Schema = StringSchema::new("First name.")
|
||||||
|
.format(&SINGLE_LINE_COMMENT_FORMAT)
|
||||||
|
.min_length(2)
|
||||||
|
.max_length(64)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const LAST_NAME_SCHEMA: Schema = StringSchema::new("Last name.")
|
||||||
|
.format(&SINGLE_LINE_COMMENT_FORMAT)
|
||||||
|
.min_length(2)
|
||||||
|
.max_length(64)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const EMAIL_SCHEMA: Schema = StringSchema::new("E-Mail Address.")
|
||||||
|
.format(&SINGLE_LINE_COMMENT_FORMAT)
|
||||||
|
.min_length(2)
|
||||||
|
.max_length(64)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
userid: {
|
||||||
|
type: Userid,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
enable: {
|
||||||
|
optional: true,
|
||||||
|
schema: ENABLE_USER_SCHEMA,
|
||||||
|
},
|
||||||
|
expire: {
|
||||||
|
optional: true,
|
||||||
|
schema: EXPIRE_USER_SCHEMA,
|
||||||
|
},
|
||||||
|
firstname: {
|
||||||
|
optional: true,
|
||||||
|
schema: FIRST_NAME_SCHEMA,
|
||||||
|
},
|
||||||
|
lastname: {
|
||||||
|
schema: LAST_NAME_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
email: {
|
||||||
|
schema: EMAIL_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
tokens: {
|
||||||
|
type: Array,
|
||||||
|
optional: true,
|
||||||
|
description: "List of user's API tokens.",
|
||||||
|
items: {
|
||||||
|
type: ApiToken
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize)]
|
||||||
|
/// User properties with added list of ApiTokens
|
||||||
|
pub struct UserWithTokens {
|
||||||
|
pub userid: Userid,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub enable: Option<bool>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub expire: Option<i64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub firstname: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub lastname: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub email: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Vec::is_empty", default)]
|
||||||
|
pub tokens: Vec<ApiToken>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
tokenid: {
|
||||||
|
schema: PROXMOX_TOKEN_ID_SCHEMA,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
enable: {
|
||||||
|
optional: true,
|
||||||
|
schema: ENABLE_USER_SCHEMA,
|
||||||
|
},
|
||||||
|
expire: {
|
||||||
|
optional: true,
|
||||||
|
schema: EXPIRE_USER_SCHEMA,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize)]
|
||||||
|
/// ApiToken properties.
|
||||||
|
pub struct ApiToken {
|
||||||
|
pub tokenid: Authid,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub enable: Option<bool>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub expire: Option<i64>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl ApiToken {
|
||||||
|
pub fn is_active(&self) -> bool {
|
||||||
|
if !self.enable.unwrap_or(true) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
if let Some(expire) = self.expire {
|
||||||
|
let now = proxmox_time::epoch_i64();
|
||||||
|
if expire > 0 && expire <= now {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api(
|
||||||
|
properties: {
|
||||||
|
userid: {
|
||||||
|
type: Userid,
|
||||||
|
},
|
||||||
|
comment: {
|
||||||
|
optional: true,
|
||||||
|
schema: SINGLE_LINE_COMMENT_SCHEMA,
|
||||||
|
},
|
||||||
|
enable: {
|
||||||
|
optional: true,
|
||||||
|
schema: ENABLE_USER_SCHEMA,
|
||||||
|
},
|
||||||
|
expire: {
|
||||||
|
optional: true,
|
||||||
|
schema: EXPIRE_USER_SCHEMA,
|
||||||
|
},
|
||||||
|
firstname: {
|
||||||
|
optional: true,
|
||||||
|
schema: FIRST_NAME_SCHEMA,
|
||||||
|
},
|
||||||
|
lastname: {
|
||||||
|
schema: LAST_NAME_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
email: {
|
||||||
|
schema: EMAIL_SCHEMA,
|
||||||
|
optional: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)]
|
||||||
|
#[derive(Serialize,Deserialize,Updater)]
|
||||||
|
/// User properties.
|
||||||
|
pub struct User {
|
||||||
|
#[updater(skip)]
|
||||||
|
pub userid: Userid,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub comment: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub enable: Option<bool>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub expire: Option<i64>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub firstname: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub lastname: Option<String>,
|
||||||
|
#[serde(skip_serializing_if="Option::is_none")]
|
||||||
|
pub email: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl User {
|
||||||
|
pub fn is_active(&self) -> bool {
|
||||||
|
if !self.enable.unwrap_or(true) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
if let Some(expire) = self.expire {
|
||||||
|
let now = proxmox_time::epoch_i64();
|
||||||
|
if expire > 0 && expire <= now {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
true
|
||||||
|
}
|
||||||
|
}
|
@ -29,19 +29,24 @@ use anyhow::{bail, format_err, Error};
|
|||||||
use lazy_static::lazy_static;
|
use lazy_static::lazy_static;
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use proxmox::api::api;
|
use proxmox_schema::{
|
||||||
use proxmox::api::schema::{ApiStringFormat, Schema, StringSchema};
|
api, const_regex, ApiStringFormat, ApiType, Schema, StringSchema, UpdaterType,
|
||||||
use proxmox::const_regex;
|
};
|
||||||
|
|
||||||
// we only allow a limited set of characters
|
// we only allow a limited set of characters
|
||||||
// colon is not allowed, because we store usernames in
|
// colon is not allowed, because we store usernames in
|
||||||
// colon separated lists)!
|
// colon separated lists)!
|
||||||
// slash is not allowed because it is used as pve API delimiter
|
// slash is not allowed because it is used as pve API delimiter
|
||||||
// also see "man useradd"
|
// also see "man useradd"
|
||||||
|
#[macro_export]
|
||||||
macro_rules! USER_NAME_REGEX_STR { () => (r"(?:[^\s:/[:cntrl:]]+)") }
|
macro_rules! USER_NAME_REGEX_STR { () => (r"(?:[^\s:/[:cntrl:]]+)") }
|
||||||
|
#[macro_export]
|
||||||
macro_rules! GROUP_NAME_REGEX_STR { () => (USER_NAME_REGEX_STR!()) }
|
macro_rules! GROUP_NAME_REGEX_STR { () => (USER_NAME_REGEX_STR!()) }
|
||||||
|
#[macro_export]
|
||||||
macro_rules! TOKEN_NAME_REGEX_STR { () => (PROXMOX_SAFE_ID_REGEX_STR!()) }
|
macro_rules! TOKEN_NAME_REGEX_STR { () => (PROXMOX_SAFE_ID_REGEX_STR!()) }
|
||||||
|
#[macro_export]
|
||||||
macro_rules! USER_ID_REGEX_STR { () => (concat!(USER_NAME_REGEX_STR!(), r"@", PROXMOX_SAFE_ID_REGEX_STR!())) }
|
macro_rules! USER_ID_REGEX_STR { () => (concat!(USER_NAME_REGEX_STR!(), r"@", PROXMOX_SAFE_ID_REGEX_STR!())) }
|
||||||
|
#[macro_export]
|
||||||
macro_rules! APITOKEN_ID_REGEX_STR { () => (concat!(USER_ID_REGEX_STR!() , r"!", TOKEN_NAME_REGEX_STR!())) }
|
macro_rules! APITOKEN_ID_REGEX_STR { () => (concat!(USER_ID_REGEX_STR!() , r"!", TOKEN_NAME_REGEX_STR!())) }
|
||||||
|
|
||||||
const_regex! {
|
const_regex! {
|
||||||
@ -93,7 +98,6 @@ pub const PROXMOX_AUTH_REALM_STRING_SCHEMA: StringSchema =
|
|||||||
.max_length(32);
|
.max_length(32);
|
||||||
pub const PROXMOX_AUTH_REALM_SCHEMA: Schema = PROXMOX_AUTH_REALM_STRING_SCHEMA.schema();
|
pub const PROXMOX_AUTH_REALM_SCHEMA: Schema = PROXMOX_AUTH_REALM_STRING_SCHEMA.schema();
|
||||||
|
|
||||||
|
|
||||||
#[api(
|
#[api(
|
||||||
type: String,
|
type: String,
|
||||||
format: &PROXMOX_USER_NAME_FORMAT,
|
format: &PROXMOX_USER_NAME_FORMAT,
|
||||||
@ -393,19 +397,21 @@ impl<'a> TryFrom<&'a str> for &'a TokennameRef {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// A complete user id consisting of a user name and a realm
|
/// A complete user id consisting of a user name and a realm
|
||||||
#[derive(Clone, Debug, PartialEq, Eq, Hash)]
|
#[derive(Clone, Debug, PartialEq, Eq, Hash, UpdaterType)]
|
||||||
pub struct Userid {
|
pub struct Userid {
|
||||||
data: String,
|
data: String,
|
||||||
name_len: usize,
|
name_len: usize,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Userid {
|
impl ApiType for Userid {
|
||||||
pub const API_SCHEMA: Schema = StringSchema::new("User ID")
|
const API_SCHEMA: Schema = StringSchema::new("User ID")
|
||||||
.format(&PROXMOX_USER_ID_FORMAT)
|
.format(&PROXMOX_USER_ID_FORMAT)
|
||||||
.min_length(3)
|
.min_length(3)
|
||||||
.max_length(64)
|
.max_length(64)
|
||||||
.schema();
|
.schema();
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Userid {
|
||||||
const fn new(data: String, name_len: usize) -> Self {
|
const fn new(data: String, name_len: usize) -> Self {
|
||||||
Self { data, name_len }
|
Self { data, name_len }
|
||||||
}
|
}
|
||||||
@ -522,19 +528,21 @@ impl PartialEq<String> for Userid {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// A complete authentication id consisting of a user id and an optional token name.
|
/// A complete authentication id consisting of a user id and an optional token name.
|
||||||
#[derive(Clone, Debug, Eq, PartialEq, Hash)]
|
#[derive(Clone, Debug, Eq, PartialEq, Hash, UpdaterType)]
|
||||||
pub struct Authid {
|
pub struct Authid {
|
||||||
user: Userid,
|
user: Userid,
|
||||||
tokenname: Option<Tokenname>
|
tokenname: Option<Tokenname>
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Authid {
|
impl ApiType for Authid {
|
||||||
pub const API_SCHEMA: Schema = StringSchema::new("Authentication ID")
|
const API_SCHEMA: Schema = StringSchema::new("Authentication ID")
|
||||||
.format(&PROXMOX_AUTH_ID_FORMAT)
|
.format(&PROXMOX_AUTH_ID_FORMAT)
|
||||||
.min_length(3)
|
.min_length(3)
|
||||||
.max_length(64)
|
.max_length(64)
|
||||||
.schema();
|
.schema();
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Authid {
|
||||||
const fn new(user: Userid, tokenname: Option<Tokenname>) -> Self {
|
const fn new(user: Userid, tokenname: Option<Tokenname>) -> Self {
|
||||||
Self { user, tokenname }
|
Self { user, tokenname }
|
||||||
}
|
}
|
79
pbs-api-types/src/zfs.rs
Normal file
79
pbs-api-types/src/zfs.rs
Normal file
@ -0,0 +1,79 @@
|
|||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use proxmox_schema::*;
|
||||||
|
|
||||||
|
const_regex! {
|
||||||
|
pub ZPOOL_NAME_REGEX = r"^[a-zA-Z][a-z0-9A-Z\-_.:]+$";
|
||||||
|
}
|
||||||
|
|
||||||
|
pub const ZFS_ASHIFT_SCHEMA: Schema = IntegerSchema::new(
|
||||||
|
"Pool sector size exponent.")
|
||||||
|
.minimum(9)
|
||||||
|
.maximum(16)
|
||||||
|
.default(12)
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
pub const ZPOOL_NAME_SCHEMA: Schema = StringSchema::new("ZFS Pool Name")
|
||||||
|
.format(&ApiStringFormat::Pattern(&ZPOOL_NAME_REGEX))
|
||||||
|
.schema();
|
||||||
|
|
||||||
|
#[api(default: "On")]
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
/// The ZFS compression algorithm to use.
|
||||||
|
pub enum ZfsCompressionType {
|
||||||
|
/// Gnu Zip
|
||||||
|
Gzip,
|
||||||
|
/// LZ4
|
||||||
|
Lz4,
|
||||||
|
/// LZJB
|
||||||
|
Lzjb,
|
||||||
|
/// ZLE
|
||||||
|
Zle,
|
||||||
|
/// ZStd
|
||||||
|
ZStd,
|
||||||
|
/// Enable compression using the default algorithm.
|
||||||
|
On,
|
||||||
|
/// Disable compression.
|
||||||
|
Off,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Copy, Clone, PartialEq, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
/// The ZFS RAID level to use.
|
||||||
|
pub enum ZfsRaidLevel {
|
||||||
|
/// Single Disk
|
||||||
|
Single,
|
||||||
|
/// Mirror
|
||||||
|
Mirror,
|
||||||
|
/// Raid10
|
||||||
|
Raid10,
|
||||||
|
/// RaidZ
|
||||||
|
RaidZ,
|
||||||
|
/// RaidZ2
|
||||||
|
RaidZ2,
|
||||||
|
/// RaidZ3
|
||||||
|
RaidZ3,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[api()]
|
||||||
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all="kebab-case")]
|
||||||
|
/// zpool list item
|
||||||
|
pub struct ZpoolListItem {
|
||||||
|
/// zpool name
|
||||||
|
pub name: String,
|
||||||
|
/// Health
|
||||||
|
pub health: String,
|
||||||
|
/// Total size
|
||||||
|
pub size: u64,
|
||||||
|
/// Used size
|
||||||
|
pub alloc: u64,
|
||||||
|
/// Free space
|
||||||
|
pub free: u64,
|
||||||
|
/// ZFS fragnentation level
|
||||||
|
pub frag: u64,
|
||||||
|
/// ZFS deduplication ratio
|
||||||
|
pub dedup: f64,
|
||||||
|
}
|
9
pbs-buildcfg/Cargo.toml
Normal file
9
pbs-buildcfg/Cargo.toml
Normal file
@ -0,0 +1,9 @@
|
|||||||
|
[package]
|
||||||
|
name = "pbs-buildcfg"
|
||||||
|
version = "2.1.2"
|
||||||
|
authors = ["Proxmox Support Team <support@proxmox.com>"]
|
||||||
|
edition = "2018"
|
||||||
|
description = "macros used for pbs related paths such as configdir and rundir"
|
||||||
|
build = "build.rs"
|
||||||
|
|
||||||
|
[dependencies]
|
24
pbs-buildcfg/build.rs
Normal file
24
pbs-buildcfg/build.rs
Normal file
@ -0,0 +1,24 @@
|
|||||||
|
// build.rs
|
||||||
|
use std::env;
|
||||||
|
use std::process::Command;
|
||||||
|
|
||||||
|
fn main() {
|
||||||
|
let repoid = match env::var("REPOID") {
|
||||||
|
Ok(repoid) => repoid,
|
||||||
|
Err(_) => {
|
||||||
|
match Command::new("git")
|
||||||
|
.args(&["rev-parse", "HEAD"])
|
||||||
|
.output()
|
||||||
|
{
|
||||||
|
Ok(output) => {
|
||||||
|
String::from_utf8(output.stdout).unwrap()
|
||||||
|
}
|
||||||
|
Err(err) => {
|
||||||
|
panic!("git rev-parse failed: {}", err);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
println!("cargo:rustc-env=REPOID={}", repoid);
|
||||||
|
}
|
@ -1,12 +1,30 @@
|
|||||||
//! Exports configuration data from the build system
|
//! Exports configuration data from the build system
|
||||||
|
|
||||||
|
pub const PROXMOX_PKG_VERSION: &str =
|
||||||
|
concat!(
|
||||||
|
env!("CARGO_PKG_VERSION_MAJOR"),
|
||||||
|
".",
|
||||||
|
env!("CARGO_PKG_VERSION_MINOR"),
|
||||||
|
);
|
||||||
|
pub const PROXMOX_PKG_RELEASE: &str = env!("CARGO_PKG_VERSION_PATCH");
|
||||||
|
pub const PROXMOX_PKG_REPOID: &str = env!("REPOID");
|
||||||
|
|
||||||
|
|
||||||
/// The configured configuration directory
|
/// The configured configuration directory
|
||||||
pub const CONFIGDIR: &str = "/etc/proxmox-backup";
|
pub const CONFIGDIR: &str = "/etc/proxmox-backup";
|
||||||
pub const JS_DIR: &str = "/usr/share/javascript/proxmox-backup";
|
pub const JS_DIR: &str = "/usr/share/javascript/proxmox-backup";
|
||||||
|
|
||||||
|
/// Unix system user used by proxmox-backup-proxy
|
||||||
|
pub const BACKUP_USER_NAME: &str = "backup";
|
||||||
|
/// Unix system group used by proxmox-backup-proxy
|
||||||
|
pub const BACKUP_GROUP_NAME: &str = "backup";
|
||||||
|
|
||||||
#[macro_export]
|
#[macro_export]
|
||||||
macro_rules! PROXMOX_BACKUP_RUN_DIR_M { () => ("/run/proxmox-backup") }
|
macro_rules! PROXMOX_BACKUP_RUN_DIR_M { () => ("/run/proxmox-backup") }
|
||||||
|
|
||||||
|
#[macro_export]
|
||||||
|
macro_rules! PROXMOX_BACKUP_STATE_DIR_M { () => ("/var/lib/proxmox-backup") }
|
||||||
|
|
||||||
#[macro_export]
|
#[macro_export]
|
||||||
macro_rules! PROXMOX_BACKUP_LOG_DIR_M { () => ("/var/log/proxmox-backup") }
|
macro_rules! PROXMOX_BACKUP_LOG_DIR_M { () => ("/var/log/proxmox-backup") }
|
||||||
|
|
||||||
@ -21,6 +39,9 @@ macro_rules! PROXMOX_BACKUP_FILE_RESTORE_BIN_DIR_M {
|
|||||||
/// namespaced directory for in-memory (tmpfs) run state
|
/// namespaced directory for in-memory (tmpfs) run state
|
||||||
pub const PROXMOX_BACKUP_RUN_DIR: &str = PROXMOX_BACKUP_RUN_DIR_M!();
|
pub const PROXMOX_BACKUP_RUN_DIR: &str = PROXMOX_BACKUP_RUN_DIR_M!();
|
||||||
|
|
||||||
|
/// namespaced directory for persistent state
|
||||||
|
pub const PROXMOX_BACKUP_STATE_DIR: &str = PROXMOX_BACKUP_STATE_DIR_M!();
|
||||||
|
|
||||||
/// namespaced directory for persistent logging
|
/// namespaced directory for persistent logging
|
||||||
pub const PROXMOX_BACKUP_LOG_DIR: &str = PROXMOX_BACKUP_LOG_DIR_M!();
|
pub const PROXMOX_BACKUP_LOG_DIR: &str = PROXMOX_BACKUP_LOG_DIR_M!();
|
||||||
|
|
||||||
@ -43,6 +64,10 @@ pub const PROXMOX_BACKUP_API_PID_FN: &str = concat!(PROXMOX_BACKUP_RUN_DIR_M!(),
|
|||||||
pub const PROXMOX_BACKUP_INITRAMFS_FN: &str =
|
pub const PROXMOX_BACKUP_INITRAMFS_FN: &str =
|
||||||
concat!(PROXMOX_BACKUP_CACHE_DIR_M!(), "/file-restore-initramfs.img");
|
concat!(PROXMOX_BACKUP_CACHE_DIR_M!(), "/file-restore-initramfs.img");
|
||||||
|
|
||||||
|
/// filename of the cached initramfs to use for debugging single file restore
|
||||||
|
pub const PROXMOX_BACKUP_INITRAMFS_DBG_FN: &str =
|
||||||
|
concat!(PROXMOX_BACKUP_CACHE_DIR_M!(), "/file-restore-initramfs-debug.img");
|
||||||
|
|
||||||
/// filename of the kernel to use for booting single file restore VMs
|
/// filename of the kernel to use for booting single file restore VMs
|
||||||
pub const PROXMOX_BACKUP_KERNEL_FN: &str =
|
pub const PROXMOX_BACKUP_KERNEL_FN: &str =
|
||||||
concat!(PROXMOX_BACKUP_FILE_RESTORE_BIN_DIR_M!(), "/bzImage");
|
concat!(PROXMOX_BACKUP_FILE_RESTORE_BIN_DIR_M!(), "/bzImage");
|
||||||
@ -52,7 +77,7 @@ pub const PROXMOX_BACKUP_KERNEL_FN: &str =
|
|||||||
/// This is a simply way to get the full path for configuration files.
|
/// This is a simply way to get the full path for configuration files.
|
||||||
/// #### Example:
|
/// #### Example:
|
||||||
/// ```
|
/// ```
|
||||||
/// # #[macro_use] extern crate proxmox_backup;
|
/// use pbs_buildcfg::configdir;
|
||||||
/// let cert_path = configdir!("/proxy.pfx");
|
/// let cert_path = configdir!("/proxy.pfx");
|
||||||
/// ```
|
/// ```
|
||||||
#[macro_export]
|
#[macro_export]
|
||||||
@ -66,6 +91,6 @@ macro_rules! configdir {
|
|||||||
#[macro_export]
|
#[macro_export]
|
||||||
macro_rules! rundir {
|
macro_rules! rundir {
|
||||||
($subdir:expr) => {
|
($subdir:expr) => {
|
||||||
concat!(PROXMOX_BACKUP_RUN_DIR_M!(), $subdir)
|
concat!($crate::PROXMOX_BACKUP_RUN_DIR_M!(), $subdir)
|
||||||
};
|
};
|
||||||
}
|
}
|
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user