Compare commits
926 Commits
v0.22.4
...
py-pycm-ne
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
53b8f91c02 | ||
|
|
a841ddd00c | ||
|
|
39455768b2 | ||
|
|
e529a454eb | ||
|
|
1b5dc396e3 | ||
|
|
15a3ac0512 | ||
|
|
52f149266f | ||
|
|
8d33c2e7c0 | ||
|
|
b3d82dc3a8 | ||
|
|
0fb44529bb | ||
|
|
6ea944bf17 | ||
|
|
8c6177c47f | ||
|
|
b65d9f1524 | ||
|
|
03e5dddf24 | ||
|
|
7bb892f7b3 | ||
|
|
66ed8ebbd9 | ||
|
|
0d326f83b6 | ||
|
|
fc0955b125 | ||
|
|
13ba1b96c3 | ||
|
|
d66d169027 | ||
|
|
6decd6aaa1 | ||
|
|
3c0ffa8652 | ||
|
|
4917e3f664 | ||
|
|
b2a14b456e | ||
|
|
ab1580a37f | ||
|
|
c1f979cd54 | ||
|
|
d124338ecb | ||
|
|
d001f14514 | ||
|
|
c43205d6de | ||
|
|
54f1af5a29 | ||
|
|
350661f027 | ||
|
|
3699df2651 | ||
|
|
63197fea3e | ||
|
|
f044194b06 | ||
|
|
29302c13e9 | ||
|
|
c4808de2ff | ||
|
|
c390a4530e | ||
|
|
be771d5d6f | ||
|
|
8b45fa089e | ||
|
|
0d04223ccd | ||
|
|
5ef222d62f | ||
|
|
6810e9ed2e | ||
|
|
a6c638d0fa | ||
|
|
fa8a512945 | ||
|
|
24b73da9e6 | ||
|
|
4447d3339c | ||
|
|
d82c9e7f2a | ||
|
|
6828a7402a | ||
|
|
a0d62a40dd | ||
|
|
712dcf6b8d | ||
|
|
ad1fc34199 | ||
|
|
ab723b25d0 | ||
|
|
016673f419 | ||
|
|
7bc6d62e9b | ||
|
|
fca9cc3e0e | ||
|
|
2a178bfbb0 | ||
|
|
3381879358 | ||
|
|
ed9058618a | ||
|
|
a4c99bad6a | ||
|
|
f31f58ff26 | ||
|
|
f84918da4b | ||
|
|
80a237e250 | ||
|
|
f52d3b26c3 | ||
|
|
2029d714a0 | ||
|
|
31ef1df74f | ||
|
|
00ae96a7cb | ||
|
|
8d2a6d6744 | ||
|
|
9443e31b1e | ||
|
|
2d8ca8af69 | ||
|
|
de4d4695c4 | ||
|
|
c8cf85223f | ||
|
|
b869538544 | ||
|
|
4710cbb281 | ||
|
|
9ae1014e55 | ||
|
|
813c0dd031 | ||
|
|
91071933d0 | ||
|
|
df5bac3e6c | ||
|
|
7b9f8abce5 | ||
|
|
a2f9d4b6a1 | ||
|
|
77e16d55c1 | ||
|
|
ecb2442566 | ||
|
|
89c0b4accf | ||
|
|
8e5b51395a | ||
|
|
c2ada0f15a | ||
|
|
6d3541c5fd | ||
|
|
31e4149067 | ||
|
|
c9fba9ec79 | ||
|
|
282627714e | ||
|
|
714dd783f9 | ||
|
|
40b390903d | ||
|
|
ce1b569b69 | ||
|
|
b539eb5aab | ||
|
|
e992e1efbd | ||
|
|
33a52dd836 | ||
|
|
b5f06fb3bc | ||
|
|
494817b616 | ||
|
|
02470a5aae | ||
|
|
42232a8ab6 | ||
|
|
cb64df45c8 | ||
|
|
a11da7bdb9 | ||
|
|
9a22ae11c6 | ||
|
|
318a7e0e30 | ||
|
|
e976f351f8 | ||
|
|
437341d40e | ||
|
|
9d7ea1a28b | ||
|
|
d85668f096 | ||
|
|
5c3a23a481 | ||
|
|
8be1f26ac6 | ||
|
|
35bd21fc64 | ||
|
|
652170fb54 | ||
|
|
d4e6c29f25 | ||
|
|
c12772e73f | ||
|
|
a26ac1dbcc | ||
|
|
2afaeba292 | ||
|
|
a14e76b98d | ||
|
|
9a3a759ed3 | ||
|
|
72f17d6961 | ||
|
|
1b967a9d98 | ||
|
|
bb954390ec | ||
|
|
bf9b6940c9 | ||
|
|
22980b9e65 | ||
|
|
483426f771 | ||
|
|
1e5b976eb7 | ||
|
|
2aa6939b96 | ||
|
|
f7e601d352 | ||
|
|
c4082931e3 | ||
|
|
cee3e5436b | ||
|
|
613fa56bfc | ||
|
|
0752d94bbf | ||
|
|
3bf1a03760 | ||
|
|
e2844e2fef | ||
|
|
2ca733bbc1 | ||
|
|
e2b6eca420 | ||
|
|
67cb19614e | ||
|
|
e464461c19 | ||
|
|
6efe88f7a1 | ||
|
|
0ce35dafe1 | ||
|
|
49e419b2df | ||
|
|
d9033d8dac | ||
|
|
517b7fb0c9 | ||
|
|
568e79a1e3 | ||
|
|
9c8846b37b | ||
|
|
737b70cbbf | ||
|
|
03d2212881 | ||
|
|
b8d10916af | ||
|
|
4fe5f35c2f | ||
|
|
cea1d8b935 | ||
|
|
e7946a3a41 | ||
|
|
5c53973220 | ||
|
|
278a38f4af | ||
|
|
39bbedf517 | ||
|
|
2153f6056d | ||
|
|
2be9b41362 | ||
|
|
f9fa024fc5 | ||
|
|
7c7ac27900 | ||
|
|
253e8b1f2a | ||
|
|
60e75c9234 | ||
|
|
fb89337b04 | ||
|
|
53f71fc4a7 | ||
|
|
12e7c1569c | ||
|
|
2eb566b884 | ||
|
|
a7444873b9 | ||
|
|
397ff11d6d | ||
|
|
285563ad01 | ||
|
|
c3111ac0b4 | ||
|
|
6505e7e02a | ||
|
|
c82889058c | ||
|
|
2f4c20567c | ||
|
|
a7d6a1188b | ||
|
|
4f5244920f | ||
|
|
feecb60b9e | ||
|
|
4f18cab8d2 | ||
|
|
e6f1b4e63a | ||
|
|
c458e985af | ||
|
|
5f234e16d0 | ||
|
|
9001e9328a | ||
|
|
b237ee3689 | ||
|
|
aa911eca40 | ||
|
|
6362c615f5 | ||
|
|
544be90469 | ||
|
|
56a1663cd9 | ||
|
|
f9a46d61fa | ||
|
|
a81451ba1f | ||
|
|
b11e370888 | ||
|
|
54ee7d4165 | ||
|
|
15efcbe042 | ||
|
|
7c5fbee327 | ||
|
|
b19c4cdcf6 | ||
|
|
3212cf86f4 | ||
|
|
fbceae7773 | ||
|
|
b921d1a920 | ||
|
|
8128b549a5 | ||
|
|
7405d95035 | ||
|
|
a04b12a3ef | ||
|
|
cbf8f2326d | ||
|
|
297874bfed | ||
|
|
74398d74ac | ||
|
|
cef9c36183 | ||
|
|
5e7430975a | ||
|
|
1456f6dba1 | ||
|
|
daf74a60ca | ||
|
|
87df95c097 | ||
|
|
9b49576875 | ||
|
|
065cbf79fc | ||
|
|
09b89e87a4 | ||
|
|
ddab6156a6 | ||
|
|
10cdfff0d1 | ||
|
|
3328416976 | ||
|
|
094a621f3c | ||
|
|
87ce5d8ccb | ||
|
|
bcdc92e25f | ||
|
|
a323fab135 | ||
|
|
d42031b075 | ||
|
|
efbb18aa25 | ||
|
|
8a430f89b3 | ||
|
|
aeaa922eef | ||
|
|
a6d5a34be3 | ||
|
|
ba79542f3c | ||
|
|
dc10c8a1ed | ||
|
|
5ab814505e | ||
|
|
1d8bdcfc04 | ||
|
|
95cf341b50 | ||
|
|
a134485b1b | ||
|
|
d39edeb9a1 | ||
|
|
453fb27be2 | ||
|
|
831f04fb7d | ||
|
|
04044a9744 | ||
|
|
8077285a63 | ||
|
|
537926c1a7 | ||
|
|
02ff3d7b1e | ||
|
|
491cb278f3 | ||
|
|
ed1ebefd8f | ||
|
|
36d64fcbd4 | ||
|
|
c5cdc2c0a2 | ||
|
|
0eca86f64f | ||
|
|
50027d76a5 | ||
|
|
b4748de5a9 | ||
|
|
8f2532c624 | ||
|
|
0726513334 | ||
|
|
f951f38883 | ||
|
|
5262412c13 | ||
|
|
f022b93249 | ||
|
|
60b5e98182 | ||
|
|
682acae9fd | ||
|
|
c0f80e9117 | ||
|
|
dcf13af459 | ||
|
|
a2e4fb6b95 | ||
|
|
9dd92f493a | ||
|
|
b23e832002 | ||
|
|
ae2f626168 | ||
|
|
a73930da81 | ||
|
|
921d446196 | ||
|
|
8e2ea5de9d | ||
|
|
fd0baca222 | ||
|
|
d2fc0d6a35 | ||
|
|
eedc9e0eaf | ||
|
|
6b85f6b405 | ||
|
|
5686a6b928 | ||
|
|
ad665c6af1 | ||
|
|
d78d6db61e | ||
|
|
f47c307bf4 | ||
|
|
5b4edb9499 | ||
|
|
a6e6093922 | ||
|
|
2e8b4e660e | ||
|
|
0ca1ee8b91 | ||
|
|
a322672259 | ||
|
|
6cab86d0c1 | ||
|
|
86e7e2e070 | ||
|
|
69fca439f4 | ||
|
|
3b90fb589f | ||
|
|
fff126204c | ||
|
|
98e626cf67 | ||
|
|
d8fe628a95 | ||
|
|
4c378840e3 | ||
|
|
18de6a480b | ||
|
|
c6da4d586b | ||
|
|
61d6fc70e8 | ||
|
|
7c65655c7e | ||
|
|
76ca264b72 | ||
|
|
aa58d3c170 | ||
|
|
a32b898a00 | ||
|
|
6fcd43ee64 | ||
|
|
f1f9f00d43 | ||
|
|
8ff27f9257 | ||
|
|
78c62532c7 | ||
|
|
a7f327dced | ||
|
|
310c435396 | ||
|
|
fa3f27e8e7 | ||
|
|
6b0fefff29 | ||
|
|
f613316282 | ||
|
|
1b5b74390f | ||
|
|
b57f88cb89 | ||
|
|
03afc2a1e6 | ||
|
|
1f6ed9324d | ||
|
|
5559772afa | ||
|
|
8728631fe0 | ||
|
|
e34d9cbe5f | ||
|
|
0efba09990 | ||
|
|
a9cb80d792 | ||
|
|
dae6fe711c | ||
|
|
c9a24bc6c5 | ||
|
|
00663f29a9 | ||
|
|
15a48990b6 | ||
|
|
af0b898c2e | ||
|
|
ddf8384bc6 | ||
|
|
670f92f42b | ||
|
|
c81b0e3d2a | ||
|
|
f56d804d85 | ||
|
|
b57f08f22b | ||
|
|
34f3b8fdd0 | ||
|
|
0b3b49b4e0 | ||
|
|
fa96422702 | ||
|
|
e12168ed24 | ||
|
|
c0f2df8e0a | ||
|
|
8807ade98f | ||
|
|
13356ddbcc | ||
|
|
974033be80 | ||
|
|
8755fc7291 | ||
|
|
17c02fe759 | ||
|
|
4c7d18a772 | ||
|
|
b28b26c39a | ||
|
|
23aed605ec | ||
|
|
fdb8d565aa | ||
|
|
9b08296236 | ||
|
|
c82d8c63fa | ||
|
|
7a8989bbfc | ||
|
|
22c86074c8 | ||
|
|
ef9e449322 | ||
|
|
6b73195478 | ||
|
|
c7b9bf6a77 | ||
|
|
a84c91b259 | ||
|
|
e6566dfd67 | ||
|
|
d6419f32b8 | ||
|
|
60ed682577 | ||
|
|
6c1fa8c30b | ||
|
|
09167fe8ac | ||
|
|
eb7951818d | ||
|
|
6959656d51 | ||
|
|
f916b50491 | ||
|
|
7160e1d3e7 | ||
|
|
16369d50a7 | ||
|
|
35fc371222 | ||
|
|
4bbf4a5e79 | ||
|
|
e5bd79b011 | ||
|
|
2267b40bda | ||
|
|
b0b6016e12 | ||
|
|
c24265fe7e | ||
|
|
9d0102ac89 | ||
|
|
52b8b3ed8d | ||
|
|
380030c59a | ||
|
|
867a813328 | ||
|
|
ded3fa50a3 | ||
|
|
84653e8d9f | ||
|
|
ac0fd7138f | ||
|
|
c0f9f47b8c | ||
|
|
7e9d24a145 | ||
|
|
99405e6a4d | ||
|
|
7500a4853c | ||
|
|
54d192e026 | ||
|
|
e6a8eba72d | ||
|
|
4d604c8c9f | ||
|
|
b081e0046f | ||
|
|
baf82c0245 | ||
|
|
884a0a392d | ||
|
|
824f2a5652 | ||
|
|
b894acf1fc | ||
|
|
536856874c | ||
|
|
8d1aaef8b8 | ||
|
|
3216c4362e | ||
|
|
2e822e65fd | ||
|
|
41fde4db8c | ||
|
|
81125c3bd8 | ||
|
|
d8b0df6f5b | ||
|
|
99e3fdb180 | ||
|
|
64cfdc07cb | ||
|
|
59fbbdd9ce | ||
|
|
adedf58297 | ||
|
|
f13ea8aa75 | ||
|
|
109a4d52b5 | ||
|
|
7f2117e2cf | ||
|
|
dfab3e8829 | ||
|
|
822622f07a | ||
|
|
c4194e4f58 | ||
|
|
05c7ff4595 | ||
|
|
a6ac78c7c6 | ||
|
|
8479122e71 | ||
|
|
3d91bfea75 | ||
|
|
27e28b33ee | ||
|
|
eaf8ac7407 | ||
|
|
1e9d550bc6 | ||
|
|
0282fe9efd | ||
|
|
ad3d9c83fe | ||
|
|
910b923c5d | ||
|
|
c1f1e1396d | ||
|
|
3139dbdd39 | ||
|
|
cc6dcdcab9 | ||
|
|
57b83e5fb2 | ||
|
|
55fe73586e | ||
|
|
3be497344a | ||
|
|
05413689b9 | ||
|
|
50da223888 | ||
|
|
719b260cf1 | ||
|
|
3848c41494 | ||
|
|
da720cafd8 | ||
|
|
8f0b029308 | ||
|
|
901f4b789d | ||
|
|
6e6fef1b0e | ||
|
|
ae131a5c7c | ||
|
|
ed59e43e1d | ||
|
|
5e8f9ed1c7 | ||
|
|
9c31ff74c4 | ||
|
|
90c8fe0182 | ||
|
|
58db81c323 | ||
|
|
e867008819 | ||
|
|
9910e06b25 | ||
|
|
a3ece7ff4d | ||
|
|
00f0ca2060 | ||
|
|
35557ac21c | ||
|
|
54662f7ae1 | ||
|
|
9cdb2a8dbb | ||
|
|
ffdab20294 | ||
|
|
e91a69a756 | ||
|
|
7665076339 | ||
|
|
49ba2d84a0 | ||
|
|
0cec923e0a | ||
|
|
b97b001dad | ||
|
|
113e231abe | ||
|
|
f43ca7a554 | ||
|
|
3c2c215619 | ||
|
|
a2b3a004bf | ||
|
|
f650133f83 | ||
|
|
81f9d5baa5 | ||
|
|
3938a85ff8 | ||
|
|
84cb604b19 | ||
|
|
093504d9a0 | ||
|
|
70a93a746d | ||
|
|
4326efddbf | ||
|
|
6f51b543f0 | ||
|
|
3316e49ad3 | ||
|
|
0c4a91cd18 | ||
|
|
07c6c3ebac | ||
|
|
e3a4a07616 | ||
|
|
57467139e5 | ||
|
|
e8bc53f37b | ||
|
|
3b78515fd4 | ||
|
|
6b052c3af9 | ||
|
|
37e2d46d7d | ||
|
|
a389eb5a08 | ||
|
|
d57f174ca3 | ||
|
|
e6ae42b1eb | ||
|
|
d911b9c48d | ||
|
|
3b35b7f4fa | ||
|
|
a82fb33b31 | ||
|
|
5f35a90529 | ||
|
|
81c620b61b | ||
|
|
12866eb0d6 | ||
|
|
24b8d0666e | ||
|
|
0f2c7248c8 | ||
|
|
8cc4ad3ac5 | ||
|
|
9ba90e322e | ||
|
|
d9c6b40d8e | ||
|
|
d792e1f052 | ||
|
|
2d464f8c89 | ||
|
|
456a8e3553 | ||
|
|
db94696cf0 | ||
|
|
72bb656b9e | ||
|
|
e092026eb8 | ||
|
|
e5f5749d67 | ||
|
|
6e4f8ea7e4 | ||
|
|
5e8eff24d2 | ||
|
|
36f1801eb8 | ||
|
|
e3deee57ba | ||
|
|
404deb99f4 | ||
|
|
f594bc7aea | ||
|
|
9ed948523a | ||
|
|
72d7c2d558 | ||
|
|
d4a7582955 | ||
|
|
42ac1f0cb2 | ||
|
|
4af61d432f | ||
|
|
52bdaa7bf5 | ||
|
|
96b42238c5 | ||
|
|
c7bd259739 | ||
|
|
0dfc360b1e | ||
|
|
5e578e2e4e | ||
|
|
93111d495b | ||
|
|
b1bbe240d7 | ||
|
|
bbb58ff4c6 | ||
|
|
7a710bee17 | ||
|
|
bc3903d0e0 | ||
|
|
61c8326180 | ||
|
|
c5caa4b838 | ||
|
|
ebdff73c8c | ||
|
|
f78beb71f7 | ||
|
|
74210c7f46 | ||
|
|
eff4451cdd | ||
|
|
8d0fc3e639 | ||
|
|
3736da3f89 | ||
|
|
221e464df3 | ||
|
|
bac5253169 | ||
|
|
fc2ee5cae8 | ||
|
|
e11f83f34b | ||
|
|
6e7fb9a308 | ||
|
|
b156a62a44 | ||
|
|
990e77c55f | ||
|
|
c2fb529819 | ||
|
|
337bf3b944 | ||
|
|
a49b2f4f16 | ||
|
|
ddcf1a4b2e | ||
|
|
82a932c078 | ||
|
|
0781615117 | ||
|
|
9151fc1653 | ||
|
|
3a83b21ce1 | ||
|
|
cfc042d901 | ||
|
|
211ad9e7d9 | ||
|
|
437b259829 | ||
|
|
f524bba869 | ||
|
|
2f31fb5f17 | ||
|
|
c3567cb199 | ||
|
|
ae4c1d11f7 | ||
|
|
cbab451c1a | ||
|
|
9cec17ca26 | ||
|
|
d9c5d91b6f | ||
|
|
6e194c6ffe | ||
|
|
8f0c28037b | ||
|
|
31aabcabf7 | ||
|
|
ca9531d205 | ||
|
|
794c5eb6a0 | ||
|
|
c6cc125e22 | ||
|
|
528c1ed9ba | ||
|
|
52cc603245 | ||
|
|
5e55af2dce | ||
|
|
24ee7c8928 | ||
|
|
605df09ae1 | ||
|
|
4aebef900c | ||
|
|
59c5bef165 | ||
|
|
a18adf74bf | ||
|
|
6426ab1b7e | ||
|
|
7d1de58378 | ||
|
|
82a54378d8 | ||
|
|
e6e8fada8b | ||
|
|
7b541ac322 | ||
|
|
b0a2ea3970 | ||
|
|
cb439a09dd | ||
|
|
f87ee334c2 | ||
|
|
e8f8cf8543 | ||
|
|
8c93fb747b | ||
|
|
1701e929bc | ||
|
|
1bb3e04263 | ||
|
|
e91ae19ec4 | ||
|
|
818ae08c61 | ||
|
|
15f32f2ca1 | ||
|
|
59aa62ea5c | ||
|
|
b4c292ddd0 | ||
|
|
a25655446a | ||
|
|
cf3d59bb2e | ||
|
|
f80287166e | ||
|
|
329dc40b98 | ||
|
|
8328c34a3e | ||
|
|
6c309d3bc9 | ||
|
|
24b49eee83 | ||
|
|
715fab340f | ||
|
|
8231e84d15 | ||
|
|
8dc91a7a5c | ||
|
|
4c195b1a06 | ||
|
|
c6fb6bf5f8 | ||
|
|
ed6161b80c | ||
|
|
93424e4565 | ||
|
|
ca00e42f1d | ||
|
|
357ee1c632 | ||
|
|
cb0a3eaade | ||
|
|
a20d34b8aa | ||
|
|
329910a620 | ||
|
|
c374d04b0d | ||
|
|
b68cf16c85 | ||
|
|
391c4cf099 | ||
|
|
8260599e98 | ||
|
|
3433c8b8a5 | ||
|
|
e53bc780e4 | ||
|
|
53346dbaa6 | ||
|
|
99994ea245 | ||
|
|
3ffe02a2fe | ||
|
|
9b77502360 | ||
|
|
96a97328cf | ||
|
|
1a400383c0 | ||
|
|
4f8ab19355 | ||
|
|
8919677faf | ||
|
|
858b185a0e | ||
|
|
bc738cea32 | ||
|
|
c2196f7d3a | ||
|
|
d45c27fdbd | ||
|
|
173084de19 | ||
|
|
fd2c5fa247 | ||
|
|
73e0e9bdff | ||
|
|
4442414d74 | ||
|
|
8dbf9005f0 | ||
|
|
0c3da1b498 | ||
|
|
278f5818b7 | ||
|
|
c2e85202c7 | ||
|
|
b021b12043 | ||
|
|
89a0c9f4b3 | ||
|
|
259629c300 | ||
|
|
1ce2baf7a2 | ||
|
|
4576a42a0f | ||
|
|
4fba351b92 | ||
|
|
706737245a | ||
|
|
0dbdf49075 | ||
|
|
641075539c | ||
|
|
9428d99991 | ||
|
|
f3cf2e94c4 | ||
|
|
85f13442d2 | ||
|
|
f478a65635 | ||
|
|
eca44600c5 | ||
|
|
d7a4652554 | ||
|
|
d85cdd1946 | ||
|
|
ce3aae1501 | ||
|
|
90c4f9d463 | ||
|
|
93ccd81d86 | ||
|
|
9c5a70ab6c | ||
|
|
5ef58144cb | ||
|
|
41ddbdfd90 | ||
|
|
66924c85a3 | ||
|
|
a4c3fc138c | ||
|
|
62ed2a07a7 | ||
|
|
e7aec9e872 | ||
|
|
b065c3e11e | ||
|
|
88b357c453 | ||
|
|
bb7299c04a | ||
|
|
7a5bddfd15 | ||
|
|
50fe769f82 | ||
|
|
29d39d1adf | ||
|
|
8dde7f3975 | ||
|
|
0cd038273e | ||
|
|
1f5bfe80ed | ||
|
|
4d2611ad8a | ||
|
|
21a97dad31 | ||
|
|
338a01ca6d | ||
|
|
392396ded4 | ||
|
|
a336e0edb7 | ||
|
|
9426fefa00 | ||
|
|
812192eef5 | ||
|
|
b8c8e80965 | ||
|
|
77fd5d8414 | ||
|
|
82050ed371 | ||
|
|
a7381a9413 | ||
|
|
b932783d4d | ||
|
|
0b51f25034 | ||
|
|
d6a182fb5d | ||
|
|
e8635adb21 | ||
|
|
f242e0fd0c | ||
|
|
67b5f6b838 | ||
|
|
9d16f17463 | ||
|
|
f44f5b0db0 | ||
|
|
39ace5fc45 | ||
|
|
0601d6a0c5 | ||
|
|
11869ff872 | ||
|
|
6753605807 | ||
|
|
918db85737 | ||
|
|
1184de8352 | ||
|
|
2470fde5d9 | ||
|
|
abfff43976 | ||
|
|
230687a501 | ||
|
|
5ff8908ff3 | ||
|
|
882e09e50b | ||
|
|
6753f4a7cb | ||
|
|
1dc63dbea6 | ||
|
|
b9dfae4722 | ||
|
|
70412612c7 | ||
|
|
cd741c368c | ||
|
|
16a7bef456 | ||
|
|
85f62728c6 | ||
|
|
092dc96e6c | ||
|
|
2bb20caa5f | ||
|
|
00bcf935e8 | ||
|
|
3751372396 | ||
|
|
e6afeca92f | ||
|
|
35b9307af6 | ||
|
|
567f728579 | ||
|
|
404c5c29a1 | ||
|
|
63712ba6c6 | ||
|
|
ef62d47dc7 | ||
|
|
a4594857fc | ||
|
|
e77572b753 | ||
|
|
8c84c5ff66 | ||
|
|
5d8beaf0ed | ||
|
|
ac405f3d79 | ||
|
|
2e30553310 | ||
|
|
85a61772d8 | ||
|
|
4007f8726d | ||
|
|
a097f7791b | ||
|
|
3d4d89b2c0 | ||
|
|
e461234865 | ||
|
|
2c1d5f9844 | ||
|
|
c4b682b983 | ||
|
|
de0b784d5a | ||
|
|
5f38afdfc7 | ||
|
|
ac67c6e34b | ||
|
|
72deb53832 | ||
|
|
7c87253fd8 | ||
|
|
1136aedd08 | ||
|
|
24e1b56268 | ||
|
|
eef6a79b35 | ||
|
|
556a36cbd7 | ||
|
|
8aa490d6b7 | ||
|
|
d9d085da10 | ||
|
|
d88d720577 | ||
|
|
1d670ae744 | ||
|
|
35ad6f52c1 | ||
|
|
b61bae7640 | ||
|
|
8b7abace8b | ||
|
|
5cf98d9564 | ||
|
|
973a961cb5 | ||
|
|
868d0cb957 | ||
|
|
497f3a3832 | ||
|
|
9843f41bce | ||
|
|
e54fefc2b7 | ||
|
|
90c0889533 | ||
|
|
6696e82ce7 | ||
|
|
dcc55d53db | ||
|
|
92000e81b8 | ||
|
|
125175ae25 | ||
|
|
f60e548a0d | ||
|
|
04dc16a6b1 | ||
|
|
27b90e38db | ||
|
|
7e5ce3ba48 | ||
|
|
f5f7cfdc8f | ||
|
|
3e1a562312 | ||
|
|
ce4d962faa | ||
|
|
b9816a97fc | ||
|
|
f7b9c30456 | ||
|
|
884620a38a | ||
|
|
7503a41773 | ||
|
|
9a5fc6b4a3 | ||
|
|
a31aeed167 | ||
|
|
71f542a951 | ||
|
|
322bd48788 | ||
|
|
b752fa59d4 | ||
|
|
d53e4cc426 | ||
|
|
ee4b7fa3a1 | ||
|
|
d6f02c86d9 | ||
|
|
62efde8e3c | ||
|
|
bda1d94d49 | ||
|
|
3f472039c5 | ||
|
|
912ef34206 | ||
|
|
9c88a48a73 | ||
|
|
4bf5cc9a9a | ||
|
|
08834e2b03 | ||
|
|
8020a111df | ||
|
|
86fb547f7c | ||
|
|
b9556c7c44 | ||
|
|
7bdb106b1b | ||
|
|
2b191cd7f4 | ||
|
|
774f0a4e60 | ||
|
|
faf11efa72 | ||
|
|
5a99142b41 | ||
|
|
a3aca0242a | ||
|
|
72f276fab3 | ||
|
|
21139945df | ||
|
|
900bd2f477 | ||
|
|
29d4a5af44 | ||
|
|
dd9b7ed6a7 | ||
|
|
09ff74be62 | ||
|
|
a94ebfea11 | ||
|
|
8f5fe1d123 | ||
|
|
d4fb58efa3 | ||
|
|
ce900346cc | ||
|
|
7cb64e465f | ||
|
|
eb70c9f5b9 | ||
|
|
a28405700e | ||
|
|
f8f4d94d7a | ||
|
|
32dfb522d6 | ||
|
|
c61c707aa5 | ||
|
|
60d10848c8 | ||
|
|
dcd6b530f9 | ||
|
|
419f0742a0 | ||
|
|
c99174798b | ||
|
|
8df2a4b511 | ||
|
|
c174cf6830 | ||
|
|
5eebd65366 | ||
|
|
625f5323c0 | ||
|
|
e05a32cead | ||
|
|
c69af5d1e5 | ||
|
|
1ac2ee8043 | ||
|
|
36af1c1c73 | ||
|
|
e2fa087002 | ||
|
|
df02bfbad2 | ||
|
|
fecb63843e | ||
|
|
b33e2d09d3 | ||
|
|
f8054aa21a | ||
|
|
8f3a2acc54 | ||
|
|
d1a20908b8 | ||
|
|
dd781f7368 | ||
|
|
9bcc43c4c1 | ||
|
|
77c83af17d | ||
|
|
574bd2db99 | ||
|
|
a76f37da96 | ||
|
|
9e75f3ec0a | ||
|
|
4d42d45897 | ||
|
|
a4b4bfda73 | ||
|
|
1bcdd3a57e | ||
|
|
297a3a1bc9 | ||
|
|
8d01e8c978 | ||
|
|
6be28aa303 | ||
|
|
5e38310515 | ||
|
|
ddfed65485 | ||
|
|
2a16d8bfa8 | ||
|
|
6a40a50a29 | ||
|
|
b2924f68c0 | ||
|
|
41ffe36636 | ||
|
|
24edc72252 | ||
|
|
83b38a26a0 | ||
|
|
914d785e3b | ||
|
|
f99f642fa8 | ||
|
|
e0bf3667e3 | ||
|
|
a24ca50fed | ||
|
|
51e9f37252 | ||
|
|
453900c884 | ||
|
|
4696459d2d | ||
|
|
ad1e3231e5 | ||
|
|
2ef7eb1826 | ||
|
|
fe86019f9a | ||
|
|
9dbb18219f | ||
|
|
451a977de0 | ||
|
|
e604929a4c | ||
|
|
9d591f9f7c | ||
|
|
f8ad915100 | ||
|
|
cbbabe6920 | ||
|
|
81fe460194 | ||
|
|
b894f996c0 | ||
|
|
1ce09847d9 | ||
|
|
722d401394 | ||
|
|
e6f04d5ef9 | ||
|
|
b8e3ecbf00 | ||
|
|
d189387c24 | ||
|
|
9e96ddc5ae | ||
|
|
543bd189af | ||
|
|
43291aa723 | ||
|
|
d0589285f7 | ||
|
|
d079aaa083 | ||
|
|
6c65977e0d | ||
|
|
1b5d786cf5 | ||
|
|
4cf00645bd | ||
|
|
e9149cfc3c | ||
|
|
a5c8111076 | ||
|
|
c3576f712d | ||
|
|
410e6a59b7 | ||
|
|
bd2b2fb75a | ||
|
|
7ae318efd0 | ||
|
|
73e9d56647 | ||
|
|
f87a752b63 | ||
|
|
ae2fec30c3 | ||
|
|
1af5564cbe | ||
|
|
a8f057a701 | ||
|
|
7f3dd38ccc | ||
|
|
8e9adefcd5 | ||
|
|
d276f9700f | ||
|
|
4f111659ec | ||
|
|
eaf330f2a8 | ||
|
|
cdaeb74dc7 | ||
|
|
fbaac46604 | ||
|
|
7f6210ee90 | ||
|
|
63f6e6079a | ||
|
|
d4fd6caae0 | ||
|
|
fd3c18b6fd | ||
|
|
725f427f25 | ||
|
|
32b3e91ef7 | ||
|
|
b7e4602268 | ||
|
|
4a98d4db93 | ||
|
|
9d6bf373be | ||
|
|
cff35c4987 | ||
|
|
d594f84b8f | ||
|
|
f8f01c336c | ||
|
|
12e3665df3 | ||
|
|
fa4778b9fc | ||
|
|
66d297d420 | ||
|
|
56251c11f3 | ||
|
|
40bf9a179b | ||
|
|
095aba0b9f | ||
|
|
4270136598 | ||
|
|
f73d7d2dce | ||
|
|
567566da08 | ||
|
|
30a9ab749d | ||
|
|
8160a96b27 | ||
|
|
10414d3e6c | ||
|
|
1d96c09094 | ||
|
|
e7112fbc6a | ||
|
|
b79761b7eb | ||
|
|
3381899c69 | ||
|
|
c7cf5eabc1 | ||
|
|
d88fa5cf8e | ||
|
|
2ed0e3d737 | ||
|
|
506a40cac1 | ||
|
|
447739fcef | ||
|
|
e60f6f4a6e | ||
|
|
7df35d0da0 | ||
|
|
71b035ece1 | ||
|
|
86a134235e | ||
|
|
24cd0da7fb | ||
|
|
762833a663 | ||
|
|
636d479e5f | ||
|
|
f2184f26fa | ||
|
|
e1686eef7c | ||
|
|
314893982e | ||
|
|
9ab6c30a3d | ||
|
|
ddf94291d4 | ||
|
|
5d1038c512 | ||
|
|
2e40c88d50 | ||
|
|
2bcba57757 | ||
|
|
37330e5e2b | ||
|
|
b4411cf2db | ||
|
|
65d1ae083c | ||
|
|
0b8faa3918 | ||
|
|
f077c7e33b | ||
|
|
9d7410d22e | ||
|
|
e295730d0e | ||
|
|
868327ee14 | ||
|
|
f5430b16bc | ||
|
|
2446695113 | ||
|
|
e0c6cca65c | ||
|
|
84ed4cd331 | ||
|
|
f6d50f790e | ||
|
|
d3c3d23d1e | ||
|
|
33752c2b55 | ||
|
|
26759249ca | ||
|
|
8b4cbbe7b3 | ||
|
|
be71f9fdc4 | ||
|
|
05c1e7ecc2 |
3
.github/dependabot.yml
vendored
3
.github/dependabot.yml
vendored
@@ -12,6 +12,7 @@ updates:
|
|||||||
interval: "daily"
|
interval: "daily"
|
||||||
# Requirements to run style checks
|
# Requirements to run style checks
|
||||||
- package-ecosystem: "pip"
|
- package-ecosystem: "pip"
|
||||||
directory: "/.github/workflows/style"
|
directories:
|
||||||
|
- "/.github/workflows/requirements/*"
|
||||||
schedule:
|
schedule:
|
||||||
interval: "daily"
|
interval: "daily"
|
||||||
|
|||||||
6
.github/workflows/audit.yaml
vendored
6
.github/workflows/audit.yaml
vendored
@@ -28,8 +28,8 @@ jobs:
|
|||||||
run:
|
run:
|
||||||
shell: ${{ matrix.system.shell }}
|
shell: ${{ matrix.system.shell }}
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: ${{inputs.python_version}}
|
python-version: ${{inputs.python_version}}
|
||||||
- name: Install Python packages
|
- name: Install Python packages
|
||||||
@@ -61,7 +61,7 @@ jobs:
|
|||||||
./share/spack/qa/validate_last_exit.ps1
|
./share/spack/qa/validate_last_exit.ps1
|
||||||
spack -d audit externals
|
spack -d audit externals
|
||||||
./share/spack/qa/validate_last_exit.ps1
|
./share/spack/qa/validate_last_exit.ps1
|
||||||
- uses: codecov/codecov-action@5ecb98a3c6b747ed38dc09f787459979aebb39be
|
- uses: codecov/codecov-action@e28ff129e5465c2c0dcc6f003fc735cb6ae0c673
|
||||||
if: ${{ inputs.with_coverage == 'true' }}
|
if: ${{ inputs.with_coverage == 'true' }}
|
||||||
with:
|
with:
|
||||||
flags: unittests,audits
|
flags: unittests,audits
|
||||||
|
|||||||
37
.github/workflows/bootstrap.yml
vendored
37
.github/workflows/bootstrap.yml
vendored
@@ -37,7 +37,7 @@ jobs:
|
|||||||
make patch unzip which xz python3 python3-devel tree \
|
make patch unzip which xz python3 python3-devel tree \
|
||||||
cmake bison
|
cmake bison
|
||||||
- name: Checkout
|
- name: Checkout
|
||||||
uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- name: Bootstrap clingo
|
- name: Bootstrap clingo
|
||||||
@@ -53,27 +53,33 @@ jobs:
|
|||||||
runs-on: ${{ matrix.runner }}
|
runs-on: ${{ matrix.runner }}
|
||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
runner: ['macos-13', 'macos-14', "ubuntu-latest"]
|
runner: ['macos-13', 'macos-14', "ubuntu-latest", "windows-latest"]
|
||||||
steps:
|
steps:
|
||||||
- name: Setup macOS
|
- name: Setup macOS
|
||||||
if: ${{ matrix.runner != 'ubuntu-latest' }}
|
if: ${{ matrix.runner != 'ubuntu-latest' && matrix.runner != 'windows-latest' }}
|
||||||
run: |
|
run: |
|
||||||
brew install cmake bison tree
|
brew install cmake bison tree
|
||||||
- name: Checkout
|
- name: Checkout
|
||||||
uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: "3.12"
|
python-version: "3.12"
|
||||||
- name: Bootstrap clingo
|
- name: Bootstrap clingo
|
||||||
|
env:
|
||||||
|
SETUP_SCRIPT_EXT: ${{ matrix.runner == 'windows-latest' && 'ps1' || 'sh' }}
|
||||||
|
SETUP_SCRIPT_SOURCE: ${{ matrix.runner == 'windows-latest' && './' || 'source ' }}
|
||||||
|
USER_SCOPE_PARENT_DIR: ${{ matrix.runner == 'windows-latest' && '$env:userprofile' || '$HOME' }}
|
||||||
|
VALIDATE_LAST_EXIT: ${{ matrix.runner == 'windows-latest' && './share/spack/qa/validate_last_exit.ps1' || '' }}
|
||||||
run: |
|
run: |
|
||||||
source share/spack/setup-env.sh
|
${{ env.SETUP_SCRIPT_SOURCE }}share/spack/setup-env.${{ env.SETUP_SCRIPT_EXT }}
|
||||||
spack bootstrap disable github-actions-v0.5
|
spack bootstrap disable github-actions-v0.5
|
||||||
spack bootstrap disable github-actions-v0.4
|
spack bootstrap disable github-actions-v0.4
|
||||||
spack external find --not-buildable cmake bison
|
spack external find --not-buildable cmake bison
|
||||||
spack -d solve zlib
|
spack -d solve zlib
|
||||||
tree ~/.spack/bootstrap/store/
|
${{ env.VALIDATE_LAST_EXIT }}
|
||||||
|
tree ${{ env.USER_SCOPE_PARENT_DIR }}/.spack/bootstrap/store/
|
||||||
|
|
||||||
gnupg-sources:
|
gnupg-sources:
|
||||||
runs-on: ${{ matrix.runner }}
|
runs-on: ${{ matrix.runner }}
|
||||||
@@ -84,15 +90,13 @@ jobs:
|
|||||||
- name: Setup macOS
|
- name: Setup macOS
|
||||||
if: ${{ matrix.runner != 'ubuntu-latest' }}
|
if: ${{ matrix.runner != 'ubuntu-latest' }}
|
||||||
run: |
|
run: |
|
||||||
brew install tree
|
brew install tree gawk
|
||||||
# Remove GnuPG since we want to bootstrap it
|
sudo rm -rf $(command -v gpg gpg2)
|
||||||
sudo rm -rf /usr/local/bin/gpg
|
|
||||||
- name: Setup Ubuntu
|
- name: Setup Ubuntu
|
||||||
if: ${{ matrix.runner == 'ubuntu-latest' }}
|
if: ${{ matrix.runner == 'ubuntu-latest' }}
|
||||||
run: |
|
run: sudo rm -rf $(command -v gpg gpg2 patchelf)
|
||||||
sudo rm -rf $(which gpg) $(which gpg2) $(which patchelf)
|
|
||||||
- name: Checkout
|
- name: Checkout
|
||||||
uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- name: Bootstrap GnuPG
|
- name: Bootstrap GnuPG
|
||||||
@@ -121,10 +125,10 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
sudo rm -rf $(which gpg) $(which gpg2) $(which patchelf)
|
sudo rm -rf $(which gpg) $(which gpg2) $(which patchelf)
|
||||||
- name: Checkout
|
- name: Checkout
|
||||||
uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: |
|
python-version: |
|
||||||
3.8
|
3.8
|
||||||
@@ -150,7 +154,7 @@ jobs:
|
|||||||
not_found=0
|
not_found=0
|
||||||
old_path="$PATH"
|
old_path="$PATH"
|
||||||
export PATH="$ver_dir:$PATH"
|
export PATH="$ver_dir:$PATH"
|
||||||
./bin/spack-tmpconfig -b ./.github/workflows/bootstrap-test.sh
|
./bin/spack-tmpconfig -b ./.github/workflows/bin/bootstrap-test.sh
|
||||||
export PATH="$old_path"
|
export PATH="$old_path"
|
||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
@@ -164,4 +168,3 @@ jobs:
|
|||||||
source share/spack/setup-env.sh
|
source share/spack/setup-env.sh
|
||||||
spack -d gpg list
|
spack -d gpg list
|
||||||
tree ~/.spack/bootstrap/store/
|
tree ~/.spack/bootstrap/store/
|
||||||
|
|
||||||
|
|||||||
21
.github/workflows/build-containers.yml
vendored
21
.github/workflows/build-containers.yml
vendored
@@ -40,8 +40,7 @@ jobs:
|
|||||||
# 1: Platforms to build for
|
# 1: Platforms to build for
|
||||||
# 2: Base image (e.g. ubuntu:22.04)
|
# 2: Base image (e.g. ubuntu:22.04)
|
||||||
dockerfile: [[amazon-linux, 'linux/amd64,linux/arm64', 'amazonlinux:2'],
|
dockerfile: [[amazon-linux, 'linux/amd64,linux/arm64', 'amazonlinux:2'],
|
||||||
[centos7, 'linux/amd64,linux/arm64,linux/ppc64le', 'centos:7'],
|
[centos-stream9, 'linux/amd64,linux/arm64,linux/ppc64le', 'centos:stream9'],
|
||||||
[centos-stream, 'linux/amd64,linux/arm64,linux/ppc64le', 'centos:stream'],
|
|
||||||
[leap15, 'linux/amd64,linux/arm64,linux/ppc64le', 'opensuse/leap:15'],
|
[leap15, 'linux/amd64,linux/arm64,linux/ppc64le', 'opensuse/leap:15'],
|
||||||
[ubuntu-focal, 'linux/amd64,linux/arm64,linux/ppc64le', 'ubuntu:20.04'],
|
[ubuntu-focal, 'linux/amd64,linux/arm64,linux/ppc64le', 'ubuntu:20.04'],
|
||||||
[ubuntu-jammy, 'linux/amd64,linux/arm64,linux/ppc64le', 'ubuntu:22.04'],
|
[ubuntu-jammy, 'linux/amd64,linux/arm64,linux/ppc64le', 'ubuntu:22.04'],
|
||||||
@@ -56,7 +55,7 @@ jobs:
|
|||||||
if: github.repository == 'spack/spack'
|
if: github.repository == 'spack/spack'
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout
|
- name: Checkout
|
||||||
uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
|
|
||||||
- uses: docker/metadata-action@8e5442c4ef9f78752691e2d8f8d19755c6f78e81
|
- uses: docker/metadata-action@8e5442c4ef9f78752691e2d8f8d19755c6f78e81
|
||||||
id: docker_meta
|
id: docker_meta
|
||||||
@@ -77,7 +76,7 @@ jobs:
|
|||||||
env:
|
env:
|
||||||
SPACK_YAML_OS: "${{ matrix.dockerfile[2] }}"
|
SPACK_YAML_OS: "${{ matrix.dockerfile[2] }}"
|
||||||
run: |
|
run: |
|
||||||
.github/workflows/generate_spack_yaml_containerize.sh
|
.github/workflows/bin/generate_spack_yaml_containerize.sh
|
||||||
. share/spack/setup-env.sh
|
. share/spack/setup-env.sh
|
||||||
mkdir -p dockerfiles/${{ matrix.dockerfile[0] }}
|
mkdir -p dockerfiles/${{ matrix.dockerfile[0] }}
|
||||||
spack containerize --last-stage=bootstrap | tee dockerfiles/${{ matrix.dockerfile[0] }}/Dockerfile
|
spack containerize --last-stage=bootstrap | tee dockerfiles/${{ matrix.dockerfile[0] }}/Dockerfile
|
||||||
@@ -88,19 +87,19 @@ jobs:
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
- name: Upload Dockerfile
|
- name: Upload Dockerfile
|
||||||
uses: actions/upload-artifact@65462800fd760344b1a7b4382951275a0abb4808
|
uses: actions/upload-artifact@0b2256b8c012f0828dc542b3febcab082c67f72b
|
||||||
with:
|
with:
|
||||||
name: dockerfiles_${{ matrix.dockerfile[0] }}
|
name: dockerfiles_${{ matrix.dockerfile[0] }}
|
||||||
path: dockerfiles
|
path: dockerfiles
|
||||||
|
|
||||||
- name: Set up QEMU
|
- name: Set up QEMU
|
||||||
uses: docker/setup-qemu-action@68827325e0b33c7199eb31dd4e31fbe9023e06e3
|
uses: docker/setup-qemu-action@49b3bc8e6bdd4a60e6116a5414239cba5943d3cf
|
||||||
|
|
||||||
- name: Set up Docker Buildx
|
- name: Set up Docker Buildx
|
||||||
uses: docker/setup-buildx-action@d70bba72b1f3fd22344832f00baa16ece964efeb
|
uses: docker/setup-buildx-action@aa33708b10e362ff993539393ff100fa93ed6a27
|
||||||
|
|
||||||
- name: Log in to GitHub Container Registry
|
- name: Log in to GitHub Container Registry
|
||||||
uses: docker/login-action@e92390c5fb421da1463c202d546fed0ec5c39f20
|
uses: docker/login-action@9780b0c442fbb1117ed29e0efdff1e18412f7567
|
||||||
with:
|
with:
|
||||||
registry: ghcr.io
|
registry: ghcr.io
|
||||||
username: ${{ github.actor }}
|
username: ${{ github.actor }}
|
||||||
@@ -108,13 +107,13 @@ jobs:
|
|||||||
|
|
||||||
- name: Log in to DockerHub
|
- name: Log in to DockerHub
|
||||||
if: github.event_name != 'pull_request'
|
if: github.event_name != 'pull_request'
|
||||||
uses: docker/login-action@e92390c5fb421da1463c202d546fed0ec5c39f20
|
uses: docker/login-action@9780b0c442fbb1117ed29e0efdff1e18412f7567
|
||||||
with:
|
with:
|
||||||
username: ${{ secrets.DOCKERHUB_USERNAME }}
|
username: ${{ secrets.DOCKERHUB_USERNAME }}
|
||||||
password: ${{ secrets.DOCKERHUB_TOKEN }}
|
password: ${{ secrets.DOCKERHUB_TOKEN }}
|
||||||
|
|
||||||
- name: Build & Deploy ${{ matrix.dockerfile[0] }}
|
- name: Build & Deploy ${{ matrix.dockerfile[0] }}
|
||||||
uses: docker/build-push-action@2cdde995de11925a030ce8070c3d77a52ffcf1c0
|
uses: docker/build-push-action@5176d81f87c23d6fc96624dfdbcd9f3830bbe445
|
||||||
with:
|
with:
|
||||||
context: dockerfiles/${{ matrix.dockerfile[0] }}
|
context: dockerfiles/${{ matrix.dockerfile[0] }}
|
||||||
platforms: ${{ matrix.dockerfile[1] }}
|
platforms: ${{ matrix.dockerfile[1] }}
|
||||||
@@ -127,7 +126,7 @@ jobs:
|
|||||||
needs: deploy-images
|
needs: deploy-images
|
||||||
steps:
|
steps:
|
||||||
- name: Merge Artifacts
|
- name: Merge Artifacts
|
||||||
uses: actions/upload-artifact/merge@65462800fd760344b1a7b4382951275a0abb4808
|
uses: actions/upload-artifact/merge@0b2256b8c012f0828dc542b3febcab082c67f72b
|
||||||
with:
|
with:
|
||||||
name: dockerfiles
|
name: dockerfiles
|
||||||
pattern: dockerfiles_*
|
pattern: dockerfiles_*
|
||||||
|
|||||||
16
.github/workflows/ci.yaml
vendored
16
.github/workflows/ci.yaml
vendored
@@ -36,7 +36,7 @@ jobs:
|
|||||||
core: ${{ steps.filter.outputs.core }}
|
core: ${{ steps.filter.outputs.core }}
|
||||||
packages: ${{ steps.filter.outputs.packages }}
|
packages: ${{ steps.filter.outputs.packages }}
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
if: ${{ github.event_name == 'push' }}
|
if: ${{ github.event_name == 'push' }}
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
@@ -53,6 +53,13 @@ jobs:
|
|||||||
- 'var/spack/repos/builtin/packages/clingo/**'
|
- 'var/spack/repos/builtin/packages/clingo/**'
|
||||||
- 'var/spack/repos/builtin/packages/python/**'
|
- 'var/spack/repos/builtin/packages/python/**'
|
||||||
- 'var/spack/repos/builtin/packages/re2c/**'
|
- 'var/spack/repos/builtin/packages/re2c/**'
|
||||||
|
- 'var/spack/repos/builtin/packages/gnupg/**'
|
||||||
|
- 'var/spack/repos/builtin/packages/libassuan/**'
|
||||||
|
- 'var/spack/repos/builtin/packages/libgcrypt/**'
|
||||||
|
- 'var/spack/repos/builtin/packages/libgpg-error/**'
|
||||||
|
- 'var/spack/repos/builtin/packages/libksba/**'
|
||||||
|
- 'var/spack/repos/builtin/packages/npth/**'
|
||||||
|
- 'var/spack/repos/builtin/packages/pinentry/**'
|
||||||
- 'lib/spack/**'
|
- 'lib/spack/**'
|
||||||
- 'share/spack/**'
|
- 'share/spack/**'
|
||||||
- '.github/workflows/bootstrap.yml'
|
- '.github/workflows/bootstrap.yml'
|
||||||
@@ -77,13 +84,8 @@ jobs:
|
|||||||
needs: [ prechecks, changes ]
|
needs: [ prechecks, changes ]
|
||||||
uses: ./.github/workflows/unit_tests.yaml
|
uses: ./.github/workflows/unit_tests.yaml
|
||||||
secrets: inherit
|
secrets: inherit
|
||||||
windows:
|
|
||||||
if: ${{ github.repository == 'spack/spack' && needs.changes.outputs.core == 'true' }}
|
|
||||||
needs: [ prechecks ]
|
|
||||||
uses: ./.github/workflows/windows_python.yml
|
|
||||||
secrets: inherit
|
|
||||||
all:
|
all:
|
||||||
needs: [ windows, unit-tests, bootstrap ]
|
needs: [ unit-tests, bootstrap ]
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- name: Success
|
- name: Success
|
||||||
|
|||||||
8
.github/workflows/install_spack.sh
vendored
8
.github/workflows/install_spack.sh
vendored
@@ -1,8 +0,0 @@
|
|||||||
#!/usr/bin/env sh
|
|
||||||
. share/spack/setup-env.sh
|
|
||||||
echo -e "config:\n build_jobs: 2" > etc/spack/config.yaml
|
|
||||||
spack config add "packages:all:target:[x86_64]"
|
|
||||||
spack compiler find
|
|
||||||
spack compiler info apple-clang
|
|
||||||
spack debug report
|
|
||||||
spack solve zlib
|
|
||||||
4
.github/workflows/nightly-win-builds.yml
vendored
4
.github/workflows/nightly-win-builds.yml
vendored
@@ -14,10 +14,10 @@ jobs:
|
|||||||
build-paraview-deps:
|
build-paraview-deps:
|
||||||
runs-on: windows-latest
|
runs-on: windows-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: 3.9
|
python-version: 3.9
|
||||||
- name: Install Python packages
|
- name: Install Python packages
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
black==24.4.2
|
black==24.4.2
|
||||||
clingo==5.7.1
|
clingo==5.7.1
|
||||||
flake8==7.0.0
|
flake8==7.1.0
|
||||||
isort==5.13.2
|
isort==5.13.2
|
||||||
mypy==1.8.0
|
mypy==1.8.0
|
||||||
types-six==1.16.21.9
|
types-six==1.16.21.20240513
|
||||||
vermin==1.6.0
|
vermin==1.6.0
|
||||||
69
.github/workflows/unit_tests.yaml
vendored
69
.github/workflows/unit_tests.yaml
vendored
@@ -51,10 +51,10 @@ jobs:
|
|||||||
on_develop: false
|
on_develop: false
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: ${{ matrix.python-version }}
|
python-version: ${{ matrix.python-version }}
|
||||||
- name: Install System packages
|
- name: Install System packages
|
||||||
@@ -72,7 +72,7 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
# Need this for the git tests to succeed.
|
# Need this for the git tests to succeed.
|
||||||
git --version
|
git --version
|
||||||
. .github/workflows/setup_git.sh
|
. .github/workflows/bin/setup_git.sh
|
||||||
- name: Bootstrap clingo
|
- name: Bootstrap clingo
|
||||||
if: ${{ matrix.concretizer == 'clingo' }}
|
if: ${{ matrix.concretizer == 'clingo' }}
|
||||||
env:
|
env:
|
||||||
@@ -91,7 +91,7 @@ jobs:
|
|||||||
UNIT_TEST_COVERAGE: ${{ matrix.python-version == '3.11' }}
|
UNIT_TEST_COVERAGE: ${{ matrix.python-version == '3.11' }}
|
||||||
run: |
|
run: |
|
||||||
share/spack/qa/run-unit-tests
|
share/spack/qa/run-unit-tests
|
||||||
- uses: codecov/codecov-action@5ecb98a3c6b747ed38dc09f787459979aebb39be
|
- uses: codecov/codecov-action@e28ff129e5465c2c0dcc6f003fc735cb6ae0c673
|
||||||
with:
|
with:
|
||||||
flags: unittests,linux,${{ matrix.concretizer }}
|
flags: unittests,linux,${{ matrix.concretizer }}
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
@@ -100,10 +100,10 @@ jobs:
|
|||||||
shell:
|
shell:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: '3.11'
|
python-version: '3.11'
|
||||||
- name: Install System packages
|
- name: Install System packages
|
||||||
@@ -118,13 +118,13 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
# Need this for the git tests to succeed.
|
# Need this for the git tests to succeed.
|
||||||
git --version
|
git --version
|
||||||
. .github/workflows/setup_git.sh
|
. .github/workflows/bin/setup_git.sh
|
||||||
- name: Run shell tests
|
- name: Run shell tests
|
||||||
env:
|
env:
|
||||||
COVERAGE: true
|
COVERAGE: true
|
||||||
run: |
|
run: |
|
||||||
share/spack/qa/run-shell-tests
|
share/spack/qa/run-shell-tests
|
||||||
- uses: codecov/codecov-action@5ecb98a3c6b747ed38dc09f787459979aebb39be
|
- uses: codecov/codecov-action@e28ff129e5465c2c0dcc6f003fc735cb6ae0c673
|
||||||
with:
|
with:
|
||||||
flags: shelltests,linux
|
flags: shelltests,linux
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
@@ -141,13 +141,13 @@ jobs:
|
|||||||
dnf install -y \
|
dnf install -y \
|
||||||
bzip2 curl file gcc-c++ gcc gcc-gfortran git gnupg2 gzip \
|
bzip2 curl file gcc-c++ gcc gcc-gfortran git gnupg2 gzip \
|
||||||
make patch tcl unzip which xz
|
make patch tcl unzip which xz
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
- name: Setup repo and non-root user
|
- name: Setup repo and non-root user
|
||||||
run: |
|
run: |
|
||||||
git --version
|
git --version
|
||||||
git config --global --add safe.directory /__w/spack/spack
|
git config --global --add safe.directory /__w/spack/spack
|
||||||
git fetch --unshallow
|
git fetch --unshallow
|
||||||
. .github/workflows/setup_git.sh
|
. .github/workflows/bin/setup_git.sh
|
||||||
useradd spack-test
|
useradd spack-test
|
||||||
chown -R spack-test .
|
chown -R spack-test .
|
||||||
- name: Run unit tests
|
- name: Run unit tests
|
||||||
@@ -160,10 +160,10 @@ jobs:
|
|||||||
clingo-cffi:
|
clingo-cffi:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: '3.11'
|
python-version: '3.11'
|
||||||
- name: Install System packages
|
- name: Install System packages
|
||||||
@@ -178,14 +178,14 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
# Need this for the git tests to succeed.
|
# Need this for the git tests to succeed.
|
||||||
git --version
|
git --version
|
||||||
. .github/workflows/setup_git.sh
|
. .github/workflows/bin/setup_git.sh
|
||||||
- name: Run unit tests (full suite with coverage)
|
- name: Run unit tests (full suite with coverage)
|
||||||
env:
|
env:
|
||||||
COVERAGE: true
|
COVERAGE: true
|
||||||
SPACK_TEST_SOLVER: clingo
|
SPACK_TEST_SOLVER: clingo
|
||||||
run: |
|
run: |
|
||||||
share/spack/qa/run-unit-tests
|
share/spack/qa/run-unit-tests
|
||||||
- uses: codecov/codecov-action@5ecb98a3c6b747ed38dc09f787459979aebb39be
|
- uses: codecov/codecov-action@e28ff129e5465c2c0dcc6f003fc735cb6ae0c673
|
||||||
with:
|
with:
|
||||||
flags: unittests,linux,clingo
|
flags: unittests,linux,clingo
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
@@ -195,13 +195,13 @@ jobs:
|
|||||||
runs-on: ${{ matrix.os }}
|
runs-on: ${{ matrix.os }}
|
||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
os: [macos-latest, macos-14]
|
os: [macos-13, macos-14]
|
||||||
python-version: ["3.11"]
|
python-version: ["3.11"]
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: ${{ matrix.python-version }}
|
python-version: ${{ matrix.python-version }}
|
||||||
- name: Install Python packages
|
- name: Install Python packages
|
||||||
@@ -217,14 +217,45 @@ jobs:
|
|||||||
SPACK_TEST_PARALLEL: 4
|
SPACK_TEST_PARALLEL: 4
|
||||||
run: |
|
run: |
|
||||||
git --version
|
git --version
|
||||||
. .github/workflows/setup_git.sh
|
. .github/workflows/bin/setup_git.sh
|
||||||
. share/spack/setup-env.sh
|
. share/spack/setup-env.sh
|
||||||
$(which spack) bootstrap disable spack-install
|
$(which spack) bootstrap disable spack-install
|
||||||
$(which spack) solve zlib
|
$(which spack) solve zlib
|
||||||
common_args=(--dist loadfile --tx '4*popen//python=./bin/spack-tmpconfig python -u ./bin/spack python' -x)
|
common_args=(--dist loadfile --tx '4*popen//python=./bin/spack-tmpconfig python -u ./bin/spack python' -x)
|
||||||
$(which spack) unit-test --verbose --cov --cov-config=pyproject.toml --cov-report=xml:coverage.xml "${common_args[@]}"
|
$(which spack) unit-test --verbose --cov --cov-config=pyproject.toml --cov-report=xml:coverage.xml "${common_args[@]}"
|
||||||
- uses: codecov/codecov-action@5ecb98a3c6b747ed38dc09f787459979aebb39be
|
- uses: codecov/codecov-action@e28ff129e5465c2c0dcc6f003fc735cb6ae0c673
|
||||||
with:
|
with:
|
||||||
flags: unittests,macos
|
flags: unittests,macos
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
verbose: true
|
verbose: true
|
||||||
|
# Run unit tests on Windows
|
||||||
|
windows:
|
||||||
|
defaults:
|
||||||
|
run:
|
||||||
|
shell:
|
||||||
|
powershell Invoke-Expression -Command "./share/spack/qa/windows_test_setup.ps1"; {0}
|
||||||
|
runs-on: windows-latest
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
|
with:
|
||||||
|
fetch-depth: 0
|
||||||
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
|
with:
|
||||||
|
python-version: 3.9
|
||||||
|
- name: Install Python packages
|
||||||
|
run: |
|
||||||
|
python -m pip install --upgrade pip pywin32 setuptools pytest-cov clingo
|
||||||
|
- name: Create local develop
|
||||||
|
run: |
|
||||||
|
./.github/workflows/bin/setup_git.ps1
|
||||||
|
- name: Unit Test
|
||||||
|
run: |
|
||||||
|
spack unit-test -x --verbose --cov --cov-config=pyproject.toml
|
||||||
|
./share/spack/qa/validate_last_exit.ps1
|
||||||
|
coverage combine -a
|
||||||
|
coverage xml
|
||||||
|
- uses: codecov/codecov-action@e28ff129e5465c2c0dcc6f003fc735cb6ae0c673
|
||||||
|
with:
|
||||||
|
flags: unittests,windows
|
||||||
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
|
verbose: true
|
||||||
|
|||||||
18
.github/workflows/valid-style.yml
vendored
18
.github/workflows/valid-style.yml
vendored
@@ -18,15 +18,15 @@ jobs:
|
|||||||
validate:
|
validate:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: '3.11'
|
python-version: '3.11'
|
||||||
cache: 'pip'
|
cache: 'pip'
|
||||||
- name: Install Python Packages
|
- name: Install Python Packages
|
||||||
run: |
|
run: |
|
||||||
pip install --upgrade pip setuptools
|
pip install --upgrade pip setuptools
|
||||||
pip install -r .github/workflows/style/requirements.txt
|
pip install -r .github/workflows/requirements/style/requirements.txt
|
||||||
- name: vermin (Spack's Core)
|
- name: vermin (Spack's Core)
|
||||||
run: vermin --backport importlib --backport argparse --violations --backport typing -t=3.6- -vvv lib/spack/spack/ lib/spack/llnl/ bin/
|
run: vermin --backport importlib --backport argparse --violations --backport typing -t=3.6- -vvv lib/spack/spack/ lib/spack/llnl/ bin/
|
||||||
- name: vermin (Repositories)
|
- name: vermin (Repositories)
|
||||||
@@ -35,22 +35,22 @@ jobs:
|
|||||||
style:
|
style:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
- uses: actions/setup-python@39cd14951b08e74b54015e9e001cdefcf80e669f
|
||||||
with:
|
with:
|
||||||
python-version: '3.11'
|
python-version: '3.11'
|
||||||
cache: 'pip'
|
cache: 'pip'
|
||||||
- name: Install Python packages
|
- name: Install Python packages
|
||||||
run: |
|
run: |
|
||||||
pip install --upgrade pip setuptools
|
pip install --upgrade pip setuptools
|
||||||
pip install -r .github/workflows/style/requirements.txt
|
pip install -r .github/workflows/requirements/style/requirements.txt
|
||||||
- name: Setup git configuration
|
- name: Setup git configuration
|
||||||
run: |
|
run: |
|
||||||
# Need this for the git tests to succeed.
|
# Need this for the git tests to succeed.
|
||||||
git --version
|
git --version
|
||||||
. .github/workflows/setup_git.sh
|
. .github/workflows/bin/setup_git.sh
|
||||||
- name: Run style tests
|
- name: Run style tests
|
||||||
run: |
|
run: |
|
||||||
share/spack/qa/run-style-tests
|
share/spack/qa/run-style-tests
|
||||||
@@ -70,13 +70,13 @@ jobs:
|
|||||||
dnf install -y \
|
dnf install -y \
|
||||||
bzip2 curl file gcc-c++ gcc gcc-gfortran git gnupg2 gzip \
|
bzip2 curl file gcc-c++ gcc gcc-gfortran git gnupg2 gzip \
|
||||||
make patch tcl unzip which xz
|
make patch tcl unzip which xz
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332
|
||||||
- name: Setup repo and non-root user
|
- name: Setup repo and non-root user
|
||||||
run: |
|
run: |
|
||||||
git --version
|
git --version
|
||||||
git config --global --add safe.directory /__w/spack/spack
|
git config --global --add safe.directory /__w/spack/spack
|
||||||
git fetch --unshallow
|
git fetch --unshallow
|
||||||
. .github/workflows/setup_git.sh
|
. .github/workflows/bin/setup_git.sh
|
||||||
useradd spack-test
|
useradd spack-test
|
||||||
chown -R spack-test .
|
chown -R spack-test .
|
||||||
- name: Bootstrap Spack development environment
|
- name: Bootstrap Spack development environment
|
||||||
|
|||||||
83
.github/workflows/windows_python.yml
vendored
83
.github/workflows/windows_python.yml
vendored
@@ -1,83 +0,0 @@
|
|||||||
name: windows
|
|
||||||
|
|
||||||
on:
|
|
||||||
workflow_call:
|
|
||||||
|
|
||||||
concurrency:
|
|
||||||
group: windows-${{github.ref}}-${{github.event.pull_request.number || github.run_number}}
|
|
||||||
cancel-in-progress: true
|
|
||||||
|
|
||||||
defaults:
|
|
||||||
run:
|
|
||||||
shell:
|
|
||||||
powershell Invoke-Expression -Command "./share/spack/qa/windows_test_setup.ps1"; {0}
|
|
||||||
jobs:
|
|
||||||
unit-tests:
|
|
||||||
runs-on: windows-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
|
||||||
with:
|
|
||||||
fetch-depth: 0
|
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
|
||||||
with:
|
|
||||||
python-version: 3.9
|
|
||||||
- name: Install Python packages
|
|
||||||
run: |
|
|
||||||
python -m pip install --upgrade pip pywin32 setuptools pytest-cov clingo
|
|
||||||
- name: Create local develop
|
|
||||||
run: |
|
|
||||||
./.github/workflows/setup_git.ps1
|
|
||||||
- name: Unit Test
|
|
||||||
run: |
|
|
||||||
spack unit-test -x --verbose --cov --cov-config=pyproject.toml --ignore=lib/spack/spack/test/cmd
|
|
||||||
./share/spack/qa/validate_last_exit.ps1
|
|
||||||
coverage combine -a
|
|
||||||
coverage xml
|
|
||||||
- uses: codecov/codecov-action@5ecb98a3c6b747ed38dc09f787459979aebb39be
|
|
||||||
with:
|
|
||||||
flags: unittests,windows
|
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
|
||||||
verbose: true
|
|
||||||
unit-tests-cmd:
|
|
||||||
runs-on: windows-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
|
||||||
with:
|
|
||||||
fetch-depth: 0
|
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
|
||||||
with:
|
|
||||||
python-version: 3.9
|
|
||||||
- name: Install Python packages
|
|
||||||
run: |
|
|
||||||
python -m pip install --upgrade pip pywin32 setuptools coverage pytest-cov clingo
|
|
||||||
- name: Create local develop
|
|
||||||
run: |
|
|
||||||
./.github/workflows/setup_git.ps1
|
|
||||||
- name: Command Unit Test
|
|
||||||
run: |
|
|
||||||
spack unit-test -x --verbose --cov --cov-config=pyproject.toml lib/spack/spack/test/cmd
|
|
||||||
./share/spack/qa/validate_last_exit.ps1
|
|
||||||
coverage combine -a
|
|
||||||
coverage xml
|
|
||||||
- uses: codecov/codecov-action@5ecb98a3c6b747ed38dc09f787459979aebb39be
|
|
||||||
with:
|
|
||||||
flags: unittests,windows
|
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
|
||||||
verbose: true
|
|
||||||
build-abseil:
|
|
||||||
runs-on: windows-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b
|
|
||||||
with:
|
|
||||||
fetch-depth: 0
|
|
||||||
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d
|
|
||||||
with:
|
|
||||||
python-version: 3.9
|
|
||||||
- name: Install Python packages
|
|
||||||
run: |
|
|
||||||
python -m pip install --upgrade pip pywin32 setuptools coverage
|
|
||||||
- name: Build Test
|
|
||||||
run: |
|
|
||||||
spack compiler find
|
|
||||||
spack -d external find cmake ninja
|
|
||||||
spack -d install abseil-cpp
|
|
||||||
366
CHANGELOG.md
366
CHANGELOG.md
@@ -1,3 +1,369 @@
|
|||||||
|
|
||||||
|
# v0.22.0 (2024-05-12)
|
||||||
|
|
||||||
|
`v0.22.0` is a major feature release.
|
||||||
|
|
||||||
|
## Features in this release
|
||||||
|
|
||||||
|
1. **Compiler dependencies**
|
||||||
|
|
||||||
|
We are in the process of making compilers proper dependencies in Spack, and a number
|
||||||
|
of changes in `v0.22` support that effort. You may notice nodes in your dependency
|
||||||
|
graphs for compiler runtime libraries like `gcc-runtime` or `libgfortran`, and you
|
||||||
|
may notice that Spack graphs now include `libc`. We've also begun moving compiler
|
||||||
|
configuration from `compilers.yaml` to `packages.yaml` to make it consistent with
|
||||||
|
other externals. We are trying to do this with the least disruption possible, so
|
||||||
|
your existing `compilers.yaml` files should still work. We expect to be done with
|
||||||
|
this transition by the `v0.23` release in November.
|
||||||
|
|
||||||
|
* #41104: Packages compiled with `%gcc` on Linux, macOS and FreeBSD now depend on a
|
||||||
|
new package `gcc-runtime`, which contains a copy of the shared compiler runtime
|
||||||
|
libraries. This enables gcc runtime libraries to be installed and relocated when
|
||||||
|
using a build cache. When building minimal Spack-generated container images it is
|
||||||
|
no longer necessary to install libgfortran, libgomp etc. using the system package
|
||||||
|
manager.
|
||||||
|
|
||||||
|
* #42062: Packages compiled with `%oneapi` now depend on a new package
|
||||||
|
`intel-oneapi-runtime`. This is similar to `gcc-runtime`, and the runtimes can
|
||||||
|
provide virtuals and compilers can inject dependencies on virtuals into compiled
|
||||||
|
packages. This allows us to model library soname compatibility and allows
|
||||||
|
compilers like `%oneapi` to provide virtuals like `sycl` (which can also be
|
||||||
|
provided by standalone libraries). Note that until we have an agreement in place
|
||||||
|
with intel, Intel packages are marked `redistribute(source=False, binary=False)`
|
||||||
|
and must be downloaded outside of Spack.
|
||||||
|
|
||||||
|
* #43272: changes to the optimization criteria of the solver improve the hit-rate of
|
||||||
|
buildcaches by a fair amount. The solver more relaxed compatibility rules and will
|
||||||
|
not try to strictly match compilers or targets of reused specs. Users can still
|
||||||
|
enforce the previous strict behavior with `require:` sections in `packages.yaml`.
|
||||||
|
Note that to enforce correct linking, Spack will *not* reuse old `%gcc` and
|
||||||
|
`%oneapi` specs that do not have the runtime libraries as a dependency.
|
||||||
|
|
||||||
|
* #43539: Spack will reuse specs built with compilers that are *not* explicitly
|
||||||
|
configured in `compilers.yaml`. Because we can now keep runtime libraries in build
|
||||||
|
cache, we do not require you to also have a local configured compiler to *use* the
|
||||||
|
runtime libraries. This improves reuse in buildcaches and avoids conflicts with OS
|
||||||
|
updates that happen underneath Spack.
|
||||||
|
|
||||||
|
* #43190: binary compatibility on `linux` is now based on the `libc` version,
|
||||||
|
instead of on the `os` tag. Spack builds now detect the host `libc` (`glibc` or
|
||||||
|
`musl`) and add it as an implicit external node in the dependency graph. Binaries
|
||||||
|
with a `libc` with the same name and a version less than or equal to that of the
|
||||||
|
detected `libc` can be reused. This is only on `linux`, not `macos` or `Windows`.
|
||||||
|
|
||||||
|
* #43464: each package that can provide a compiler is now detectable using `spack
|
||||||
|
external find`. External packages defining compiler paths are effectively used as
|
||||||
|
compilers, and `spack external find -t compiler` can be used as a substitute for
|
||||||
|
`spack compiler find`. More details on this transition are in
|
||||||
|
[the docs](https://spack.readthedocs.io/en/latest/getting_started.html#manual-compiler-configuration)
|
||||||
|
|
||||||
|
2. **Improved `spack find` UI for Environments**
|
||||||
|
|
||||||
|
If you're working in an enviroment, you likely care about:
|
||||||
|
|
||||||
|
* What are the roots
|
||||||
|
* Which ones are installed / not installed
|
||||||
|
* What's been added that still needs to be concretized
|
||||||
|
|
||||||
|
We've tweaked `spack find` in environments to show this information much more
|
||||||
|
clearly. Installation status is shown next to each root, so you can see what is
|
||||||
|
installed. Roots are also shown in bold in the list of installed packages. There is
|
||||||
|
also a new option for `spack find -r` / `--only-roots` that will only show env
|
||||||
|
roots, if you don't want to look at all the installed specs.
|
||||||
|
|
||||||
|
More details in #42334.
|
||||||
|
|
||||||
|
3. **Improved command-line string quoting**
|
||||||
|
|
||||||
|
We are making some breaking changes to how Spack parses specs on the CLI in order to
|
||||||
|
respect shell quoting instead of trying to fight it. If you (sadly) had to write
|
||||||
|
something like this on the command line:
|
||||||
|
|
||||||
|
```
|
||||||
|
spack install zlib cflags=\"-O2 -g\"
|
||||||
|
```
|
||||||
|
|
||||||
|
That will now result in an error, but you can now write what you probably expected
|
||||||
|
to work in the first place:
|
||||||
|
|
||||||
|
```
|
||||||
|
spack install zlib cflags="-O2 -g"
|
||||||
|
```
|
||||||
|
|
||||||
|
Quoted can also now include special characters, so you can supply flags like:
|
||||||
|
|
||||||
|
```
|
||||||
|
spack intall zlib ldflags='-Wl,-rpath=$ORIGIN/_libs'
|
||||||
|
```
|
||||||
|
|
||||||
|
To reduce ambiguity in parsing, we now require that you *not* put spaces around `=`
|
||||||
|
and `==` when for flags or variants. This would not have broken before but will now
|
||||||
|
result in an error:
|
||||||
|
|
||||||
|
```
|
||||||
|
spack install zlib cflags = "-O2 -g"
|
||||||
|
```
|
||||||
|
|
||||||
|
More details and discussion in #30634.
|
||||||
|
|
||||||
|
4. **Revert default `spack install` behavior to `--reuse`**
|
||||||
|
|
||||||
|
We changed the default concretizer behavior from `--reuse` to `--reuse-deps` in
|
||||||
|
#30990 (in `v0.20`), which meant that *every* `spack install` invocation would
|
||||||
|
attempt to build a new version of the requested package / any environment roots.
|
||||||
|
While this is a common ask for *upgrading* and for *developer* workflows, we don't
|
||||||
|
think it should be the default for a package manager.
|
||||||
|
|
||||||
|
We are going to try to stick to this policy:
|
||||||
|
1. Prioritize reuse and build as little as possible by default.
|
||||||
|
2. Only upgrade or install duplicates if they are explicitly asked for, or if there
|
||||||
|
is a known security issue that necessitates an upgrade.
|
||||||
|
|
||||||
|
With the install command you now have three options:
|
||||||
|
|
||||||
|
* `--reuse` (default): reuse as many existing installations as possible.
|
||||||
|
* `--reuse-deps` / `--fresh-roots`: upgrade (freshen) roots but reuse dependencies if possible.
|
||||||
|
* `--fresh`: install fresh versions of requested packages (roots) and their dependencies.
|
||||||
|
|
||||||
|
We've also introduced `--fresh-roots` as an alias for `--reuse-deps` to make it more clear
|
||||||
|
that it may give you fresh versions. More details in #41302 and #43988.
|
||||||
|
|
||||||
|
5. **More control over reused specs**
|
||||||
|
|
||||||
|
You can now control which packages to reuse and how. There is a new
|
||||||
|
`concretizer:reuse` config option, which accepts the following properties:
|
||||||
|
|
||||||
|
- `roots`: `true` to reuse roots, `false` to reuse just dependencies
|
||||||
|
- `exclude`: list of constraints used to select which specs *not* to reuse
|
||||||
|
- `include`: list of constraints used to select which specs *to* reuse
|
||||||
|
- `from`: list of sources for reused specs (some combination of `local`,
|
||||||
|
`buildcache`, or `external`)
|
||||||
|
|
||||||
|
For example, to reuse only specs compiled with GCC, you could write:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
concretizer:
|
||||||
|
reuse:
|
||||||
|
roots: true
|
||||||
|
include:
|
||||||
|
- "%gcc"
|
||||||
|
```
|
||||||
|
|
||||||
|
Or, if `openmpi` must be used from externals, and it must be the only external used:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
concretizer:
|
||||||
|
reuse:
|
||||||
|
roots: true
|
||||||
|
from:
|
||||||
|
- type: local
|
||||||
|
exclude: ["openmpi"]
|
||||||
|
- type: buildcache
|
||||||
|
exclude: ["openmpi"]
|
||||||
|
- type: external
|
||||||
|
include: ["openmpi"]
|
||||||
|
```
|
||||||
|
|
||||||
|
6. **New `redistribute()` directive**
|
||||||
|
|
||||||
|
Some packages can't be redistributed in source or binary form. We need an explicit
|
||||||
|
way to say that in a package.
|
||||||
|
|
||||||
|
Now there is a `redistribute()` directive so that package authors can write:
|
||||||
|
|
||||||
|
```python
|
||||||
|
class MyPackage(Package):
|
||||||
|
redistribute(source=False, binary=False)
|
||||||
|
```
|
||||||
|
|
||||||
|
Like other directives, this works with `when=`:
|
||||||
|
|
||||||
|
```python
|
||||||
|
class MyPackage(Package):
|
||||||
|
# 12.0 and higher are proprietary
|
||||||
|
redistribute(source=False, binary=False, when="@12.0:")
|
||||||
|
|
||||||
|
# can't redistribute when we depend on some proprietary dependency
|
||||||
|
redistribute(source=False, binary=False, when="^proprietary-dependency")
|
||||||
|
```
|
||||||
|
|
||||||
|
More in #20185.
|
||||||
|
|
||||||
|
7. **New `conflict:` and `prefer:` syntax for package preferences**
|
||||||
|
|
||||||
|
Previously, you could express conflicts and preferences in `packages.yaml` through
|
||||||
|
some contortions with `require:`:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
packages:
|
||||||
|
zlib-ng:
|
||||||
|
require:
|
||||||
|
- one_of: ["%clang", "@:"] # conflict on %clang
|
||||||
|
- any_of: ["+shared", "@:"] # strong preference for +shared
|
||||||
|
```
|
||||||
|
|
||||||
|
You can now use `require:` and `prefer:` for a much more readable configuration:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
packages:
|
||||||
|
zlib-ng:
|
||||||
|
conflict:
|
||||||
|
- "%clang"
|
||||||
|
prefer:
|
||||||
|
- "+shared"
|
||||||
|
```
|
||||||
|
|
||||||
|
See [the documentation](https://spack.readthedocs.io/en/latest/packages_yaml.html#conflicts-and-strong-preferences)
|
||||||
|
and #41832 for more details.
|
||||||
|
|
||||||
|
8. **`include_concrete` in environments**
|
||||||
|
|
||||||
|
You may want to build on the *concrete* contents of another environment without
|
||||||
|
changing that environment. You can now include the concrete specs from another
|
||||||
|
environment's `spack.lock` with `include_concrete`:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
spack:
|
||||||
|
specs: []
|
||||||
|
concretizer:
|
||||||
|
unify: true
|
||||||
|
include_concrete:
|
||||||
|
- /path/to/environment1
|
||||||
|
- /path/to/environment2
|
||||||
|
```
|
||||||
|
|
||||||
|
Now, when *this* environment is concretized, it will bring in the already concrete
|
||||||
|
specs from `environment1` and `environment2`, and build on top of them without
|
||||||
|
changing them. This is useful if you have phased deployments, where old deployments
|
||||||
|
should not be modified but you want to use as many of them as possible. More details
|
||||||
|
in #33768.
|
||||||
|
|
||||||
|
9. **`python-venv` isolation**
|
||||||
|
|
||||||
|
Spack has unique requirements for Python because it:
|
||||||
|
1. installs every package in its own independent directory, and
|
||||||
|
2. allows users to register *external* python installations.
|
||||||
|
|
||||||
|
External installations may contain their own installed packages that can interfere
|
||||||
|
with Spack installations, and some distributions (Debian and Ubuntu) even change the
|
||||||
|
`sysconfig` in ways that alter the installation layout of installed Python packages
|
||||||
|
(e.g., with the addition of a `/local` prefix on Debian or Ubuntu). To isolate Spack
|
||||||
|
from these and other issues, we now insert a small `python-venv` package in between
|
||||||
|
`python` and packages that need to install Python code. This isolates Spack's build
|
||||||
|
environment, isolates Spack from any issues with an external python, and resolves a
|
||||||
|
large number of issues we've had with Python installations.
|
||||||
|
|
||||||
|
See #40773 for further details.
|
||||||
|
|
||||||
|
## New commands, options, and directives
|
||||||
|
|
||||||
|
* Allow packages to be pushed to build cache after install from source (#42423)
|
||||||
|
* `spack develop`: stage build artifacts in same root as non-dev builds #41373
|
||||||
|
* Don't delete `spack develop` build artifacts after install (#43424)
|
||||||
|
* `spack find`: add options for local/upstream only (#42999)
|
||||||
|
* `spack logs`: print log files for packages (either partially built or installed) (#42202)
|
||||||
|
* `patch`: support reversing patches (#43040)
|
||||||
|
* `develop`: Add -b/--build-directory option to set build_directory package attribute (#39606)
|
||||||
|
* `spack list`: add `--namesapce` / `--repo` option (#41948)
|
||||||
|
* directives: add `checked_by` field to `license()`, add some license checks
|
||||||
|
* `spack gc`: add options for environments and build dependencies (#41731)
|
||||||
|
* Add `--create` to `spack env activate` (#40896)
|
||||||
|
|
||||||
|
## Performance improvements
|
||||||
|
|
||||||
|
* environment.py: fix excessive re-reads (#43746)
|
||||||
|
* ruamel yaml: fix quadratic complexity bug (#43745)
|
||||||
|
* Refactor to improve `spec format` speed (#43712)
|
||||||
|
* Do not acquire a write lock on the env post install if no views (#43505)
|
||||||
|
* asp.py: fewer calls to `spec.copy()` (#43715)
|
||||||
|
* spec.py: early return in `__str__`
|
||||||
|
* avoid `jinja2` import at startup unless needed (#43237)
|
||||||
|
|
||||||
|
## Other new features of note
|
||||||
|
|
||||||
|
* `archspec`: update to `v0.2.4`: support for Windows, bugfixes for `neoverse-v1` and
|
||||||
|
`neoverse-v2` detection.
|
||||||
|
* `spack config get`/`blame`: with no args, show entire config
|
||||||
|
* `spack env create <env>`: dir if dir-like (#44024)
|
||||||
|
* ASP-based solver: update os compatibility for macOS (#43862)
|
||||||
|
* Add handling of custom ssl certs in urllib ops (#42953)
|
||||||
|
* Add ability to rename environments (#43296)
|
||||||
|
* Add config option and compiler support to reuse across OS's (#42693)
|
||||||
|
* Support for prereleases (#43140)
|
||||||
|
* Only reuse externals when configured (#41707)
|
||||||
|
* Environments: Add support for including views (#42250)
|
||||||
|
|
||||||
|
## Binary caches
|
||||||
|
* Build cache: make signed/unsigned a mirror property (#41507)
|
||||||
|
* tools stack
|
||||||
|
|
||||||
|
## Removals, deprecations, and syntax changes
|
||||||
|
* remove `dpcpp` compiler and package (#43418)
|
||||||
|
* spack load: remove --only argument (#42120)
|
||||||
|
|
||||||
|
## Notable Bugfixes
|
||||||
|
* repo.py: drop deleted packages from provider cache (#43779)
|
||||||
|
* Allow `+` in module file names (#41999)
|
||||||
|
* `cmd/python`: use runpy to allow multiprocessing in scripts (#41789)
|
||||||
|
* Show extension commands with spack -h (#41726)
|
||||||
|
* Support environment variable expansion inside module projections (#42917)
|
||||||
|
* Alert user to failed concretizations (#42655)
|
||||||
|
* shell: fix zsh color formatting for PS1 in environments (#39497)
|
||||||
|
* spack mirror create --all: include patches (#41579)
|
||||||
|
|
||||||
|
## Spack community stats
|
||||||
|
|
||||||
|
* 7,994 total packages; 525 since `v0.21.0`
|
||||||
|
* 178 new Python packages, 5 new R packages
|
||||||
|
* 358 people contributed to this release
|
||||||
|
* 344 committers to packages
|
||||||
|
* 45 committers to core
|
||||||
|
|
||||||
|
# v0.21.2 (2024-03-01)
|
||||||
|
|
||||||
|
## Bugfixes
|
||||||
|
|
||||||
|
- Containerize: accommodate nested or pre-existing spack-env paths (#41558)
|
||||||
|
- Fix setup-env script, when going back and forth between instances (#40924)
|
||||||
|
- Fix using fully-qualified namespaces from root specs (#41957)
|
||||||
|
- Fix a bug when a required provider is requested for multiple virtuals (#42088)
|
||||||
|
- OCI buildcaches:
|
||||||
|
- only push in parallel when forking (#42143)
|
||||||
|
- use pickleable errors (#42160)
|
||||||
|
- Fix using sticky variants in externals (#42253)
|
||||||
|
- Fix a rare issue with conditional requirements and multi-valued variants (#42566)
|
||||||
|
|
||||||
|
## Package updates
|
||||||
|
- rust: add v1.75, rework a few variants (#41161,#41903)
|
||||||
|
- py-transformers: add v4.35.2 (#41266)
|
||||||
|
- mgard: fix OpenMP on AppleClang (#42933)
|
||||||
|
|
||||||
|
# v0.21.1 (2024-01-11)
|
||||||
|
|
||||||
|
## New features
|
||||||
|
- Add support for reading buildcaches created by Spack v0.22 (#41773)
|
||||||
|
|
||||||
|
## Bugfixes
|
||||||
|
|
||||||
|
- spack graph: fix coloring with environments (#41240)
|
||||||
|
- spack info: sort variants in --variants-by-name (#41389)
|
||||||
|
- Spec.format: error on old style format strings (#41934)
|
||||||
|
- ASP-based solver:
|
||||||
|
- fix infinite recursion when computing concretization errors (#41061)
|
||||||
|
- don't error for type mismatch on preferences (#41138)
|
||||||
|
- don't emit spurious debug output (#41218)
|
||||||
|
- Improve the error message for deprecated preferences (#41075)
|
||||||
|
- Fix MSVC preview version breaking clingo build on Windows (#41185)
|
||||||
|
- Fix multi-word aliases (#41126)
|
||||||
|
- Add a warning for unconfigured compiler (#41213)
|
||||||
|
- environment: fix an issue with deconcretization/reconcretization of specs (#41294)
|
||||||
|
- buildcache: don't error if a patch is missing, when installing from binaries (#41986)
|
||||||
|
- Multiple improvements to unit-tests (#41215,#41369,#41495,#41359,#41361,#41345,#41342,#41308,#41226)
|
||||||
|
|
||||||
|
## Package updates
|
||||||
|
- root: add a webgui patch to address security issue (#41404)
|
||||||
|
- BerkeleyGW: update source urls (#38218)
|
||||||
|
|
||||||
# v0.21.0 (2023-11-11)
|
# v0.21.0 (2023-11-11)
|
||||||
|
|
||||||
`v0.21.0` is a major feature release.
|
`v0.21.0` is a major feature release.
|
||||||
|
|||||||
@@ -32,7 +32,7 @@
|
|||||||
|
|
||||||
Spack is a multi-platform package manager that builds and installs
|
Spack is a multi-platform package manager that builds and installs
|
||||||
multiple versions and configurations of software. It works on Linux,
|
multiple versions and configurations of software. It works on Linux,
|
||||||
macOS, and many supercomputers. Spack is non-destructive: installing a
|
macOS, Windows, and many supercomputers. Spack is non-destructive: installing a
|
||||||
new version of a package does not break existing installations, so many
|
new version of a package does not break existing installations, so many
|
||||||
configurations of the same package can coexist.
|
configurations of the same package can coexist.
|
||||||
|
|
||||||
|
|||||||
@@ -22,4 +22,4 @@
|
|||||||
#
|
#
|
||||||
# This is compatible across platforms.
|
# This is compatible across platforms.
|
||||||
#
|
#
|
||||||
exec /usr/bin/env spack python "$@"
|
exec spack python "$@"
|
||||||
|
|||||||
@@ -188,25 +188,27 @@ if NOT "%_sp_args%"=="%_sp_args:--help=%" (
|
|||||||
goto :end_switch
|
goto :end_switch
|
||||||
|
|
||||||
:case_load
|
:case_load
|
||||||
:: If args contain --sh, --csh, or -h/--help: just execute.
|
if NOT defined _sp_args (
|
||||||
if defined _sp_args (
|
exit /B 0
|
||||||
if NOT "%_sp_args%"=="%_sp_args:--help=%" (
|
)
|
||||||
goto :default_case
|
|
||||||
) else if NOT "%_sp_args%"=="%_sp_args:-h=%" (
|
:: If args contain --bat, or -h/--help: just execute.
|
||||||
goto :default_case
|
if NOT "%_sp_args%"=="%_sp_args:--help=%" (
|
||||||
) else if NOT "%_sp_args%"=="%_sp_args:--bat=%" (
|
goto :default_case
|
||||||
goto :default_case
|
) else if NOT "%_sp_args%"=="%_sp_args:-h=%" (
|
||||||
)
|
goto :default_case
|
||||||
|
) else if NOT "%_sp_args%"=="%_sp_args:--bat=%" (
|
||||||
|
goto :default_case
|
||||||
|
) else if NOT "%_sp_args%"=="%_sp_args:--list=%" (
|
||||||
|
goto :default_case
|
||||||
)
|
)
|
||||||
|
|
||||||
for /f "tokens=* USEBACKQ" %%I in (
|
for /f "tokens=* USEBACKQ" %%I in (
|
||||||
`python "%spack%" %_sp_flags% %_sp_subcommand% --bat %_sp_args%`) do %%I
|
`python "%spack%" %_sp_flags% %_sp_subcommand% --bat %_sp_args%`
|
||||||
|
) do %%I
|
||||||
|
|
||||||
goto :end_switch
|
goto :end_switch
|
||||||
|
|
||||||
:case_unload
|
|
||||||
goto :case_load
|
|
||||||
|
|
||||||
:default_case
|
:default_case
|
||||||
python "%spack%" %_sp_flags% %_sp_subcommand% %_sp_args%
|
python "%spack%" %_sp_flags% %_sp_subcommand% %_sp_args%
|
||||||
goto :end_switch
|
goto :end_switch
|
||||||
|
|||||||
@@ -144,3 +144,5 @@ switch($SpackSubCommand)
|
|||||||
"unload" {Invoke-SpackLoad}
|
"unload" {Invoke-SpackLoad}
|
||||||
default {python "$Env:SPACK_ROOT/bin/spack" $SpackCMD_params $SpackSubCommand $SpackSubCommandArgs}
|
default {python "$Env:SPACK_ROOT/bin/spack" $SpackCMD_params $SpackSubCommand $SpackSubCommandArgs}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
exit $LASTEXITCODE
|
||||||
|
|||||||
@@ -1,16 +0,0 @@
|
|||||||
# -------------------------------------------------------------------------
|
|
||||||
# This is the default configuration for Spack's module file generation.
|
|
||||||
#
|
|
||||||
# Settings here are versioned with Spack and are intended to provide
|
|
||||||
# sensible defaults out of the box. Spack maintainers should edit this
|
|
||||||
# file to keep it current.
|
|
||||||
#
|
|
||||||
# Users can override these settings by editing the following files.
|
|
||||||
#
|
|
||||||
# Per-spack-instance settings (overrides defaults):
|
|
||||||
# $SPACK_ROOT/etc/spack/modules.yaml
|
|
||||||
#
|
|
||||||
# Per-user settings (overrides default and site settings):
|
|
||||||
# ~/.spack/modules.yaml
|
|
||||||
# -------------------------------------------------------------------------
|
|
||||||
modules: {}
|
|
||||||
@@ -1,19 +0,0 @@
|
|||||||
# -------------------------------------------------------------------------
|
|
||||||
# This file controls default concretization preferences for Spack.
|
|
||||||
#
|
|
||||||
# Settings here are versioned with Spack and are intended to provide
|
|
||||||
# sensible defaults out of the box. Spack maintainers should edit this
|
|
||||||
# file to keep it current.
|
|
||||||
#
|
|
||||||
# Users can override these settings by editing the following files.
|
|
||||||
#
|
|
||||||
# Per-spack-instance settings (overrides defaults):
|
|
||||||
# $SPACK_ROOT/etc/spack/packages.yaml
|
|
||||||
#
|
|
||||||
# Per-user settings (overrides default and site settings):
|
|
||||||
# ~/.spack/packages.yaml
|
|
||||||
# -------------------------------------------------------------------------
|
|
||||||
packages:
|
|
||||||
all:
|
|
||||||
providers:
|
|
||||||
iconv: [glibc, musl, libiconv]
|
|
||||||
@@ -1,19 +0,0 @@
|
|||||||
# -------------------------------------------------------------------------
|
|
||||||
# This file controls default concretization preferences for Spack.
|
|
||||||
#
|
|
||||||
# Settings here are versioned with Spack and are intended to provide
|
|
||||||
# sensible defaults out of the box. Spack maintainers should edit this
|
|
||||||
# file to keep it current.
|
|
||||||
#
|
|
||||||
# Users can override these settings by editing the following files.
|
|
||||||
#
|
|
||||||
# Per-spack-instance settings (overrides defaults):
|
|
||||||
# $SPACK_ROOT/etc/spack/packages.yaml
|
|
||||||
#
|
|
||||||
# Per-user settings (overrides default and site settings):
|
|
||||||
# ~/.spack/packages.yaml
|
|
||||||
# -------------------------------------------------------------------------
|
|
||||||
packages:
|
|
||||||
all:
|
|
||||||
providers:
|
|
||||||
iconv: [glibc, musl, libiconv]
|
|
||||||
@@ -38,10 +38,9 @@ packages:
|
|||||||
lapack: [openblas, amdlibflame]
|
lapack: [openblas, amdlibflame]
|
||||||
libc: [glibc, musl]
|
libc: [glibc, musl]
|
||||||
libgfortran: [ gcc-runtime ]
|
libgfortran: [ gcc-runtime ]
|
||||||
libglx: [mesa+glx, mesa18+glx]
|
libglx: [mesa+glx]
|
||||||
libifcore: [ intel-oneapi-runtime ]
|
libifcore: [ intel-oneapi-runtime ]
|
||||||
libllvm: [llvm]
|
libllvm: [llvm]
|
||||||
libosmesa: [mesa+osmesa, mesa18+osmesa]
|
|
||||||
lua-lang: [lua, lua-luajit-openresty, lua-luajit]
|
lua-lang: [lua, lua-luajit-openresty, lua-luajit]
|
||||||
luajit: [lua-luajit-openresty, lua-luajit]
|
luajit: [lua-luajit-openresty, lua-luajit]
|
||||||
mariadb-client: [mariadb-c-client, mariadb]
|
mariadb-client: [mariadb-c-client, mariadb]
|
||||||
|
|||||||
@@ -1433,22 +1433,12 @@ the reserved keywords ``platform``, ``os`` and ``target``:
|
|||||||
$ spack install libelf os=ubuntu18.04
|
$ spack install libelf os=ubuntu18.04
|
||||||
$ spack install libelf target=broadwell
|
$ spack install libelf target=broadwell
|
||||||
|
|
||||||
or together by using the reserved keyword ``arch``:
|
|
||||||
|
|
||||||
.. code-block:: console
|
|
||||||
|
|
||||||
$ spack install libelf arch=cray-CNL10-haswell
|
|
||||||
|
|
||||||
Normally users don't have to bother specifying the architecture if they
|
Normally users don't have to bother specifying the architecture if they
|
||||||
are installing software for their current host, as in that case the
|
are installing software for their current host, as in that case the
|
||||||
values will be detected automatically. If you need fine-grained control
|
values will be detected automatically. If you need fine-grained control
|
||||||
over which packages use which targets (or over *all* packages' default
|
over which packages use which targets (or over *all* packages' default
|
||||||
target), see :ref:`package-preferences`.
|
target), see :ref:`package-preferences`.
|
||||||
|
|
||||||
.. admonition:: Cray machines
|
|
||||||
|
|
||||||
The situation is a little bit different for Cray machines and a detailed
|
|
||||||
explanation on how the architecture can be set on them can be found at :ref:`cray-support`
|
|
||||||
|
|
||||||
.. _support-for-microarchitectures:
|
.. _support-for-microarchitectures:
|
||||||
|
|
||||||
|
|||||||
@@ -147,6 +147,15 @@ example, the ``bash`` shell is used to run the ``autogen.sh`` script.
|
|||||||
def autoreconf(self, spec, prefix):
|
def autoreconf(self, spec, prefix):
|
||||||
which("bash")("autogen.sh")
|
which("bash")("autogen.sh")
|
||||||
|
|
||||||
|
If the ``package.py`` has build instructions in a separate
|
||||||
|
:ref:`builder class <multiple_build_systems>`, the signature for a phase changes slightly:
|
||||||
|
|
||||||
|
.. code-block:: python
|
||||||
|
|
||||||
|
class AutotoolsBuilder(AutotoolsBuilder):
|
||||||
|
def autoreconf(self, pkg, spec, prefix):
|
||||||
|
which("bash")("autogen.sh")
|
||||||
|
|
||||||
"""""""""""""""""""""""""""""""""""""""
|
"""""""""""""""""""""""""""""""""""""""
|
||||||
patching configure or Makefile.in files
|
patching configure or Makefile.in files
|
||||||
"""""""""""""""""""""""""""""""""""""""
|
"""""""""""""""""""""""""""""""""""""""
|
||||||
|
|||||||
@@ -25,7 +25,7 @@ use Spack to build packages with the tools.
|
|||||||
The Spack Python class ``IntelOneapiPackage`` is a base class that is
|
The Spack Python class ``IntelOneapiPackage`` is a base class that is
|
||||||
used by ``IntelOneapiCompilers``, ``IntelOneapiMkl``,
|
used by ``IntelOneapiCompilers``, ``IntelOneapiMkl``,
|
||||||
``IntelOneapiTbb`` and other classes to implement the oneAPI
|
``IntelOneapiTbb`` and other classes to implement the oneAPI
|
||||||
packages. Search for ``oneAPI`` at `<packages.spack.io>`_ for the full
|
packages. Search for ``oneAPI`` at `packages.spack.io <https://packages.spack.io>`_ for the full
|
||||||
list of available oneAPI packages, or use::
|
list of available oneAPI packages, or use::
|
||||||
|
|
||||||
spack list -d oneAPI
|
spack list -d oneAPI
|
||||||
|
|||||||
@@ -11,7 +11,8 @@ Chaining Spack Installations
|
|||||||
|
|
||||||
You can point your Spack installation to another installation to use any
|
You can point your Spack installation to another installation to use any
|
||||||
packages that are installed there. To register the other Spack instance,
|
packages that are installed there. To register the other Spack instance,
|
||||||
you can add it as an entry to ``upstreams.yaml``:
|
you can add it as an entry to ``upstreams.yaml`` at any of the
|
||||||
|
:ref:`configuration-scopes`:
|
||||||
|
|
||||||
.. code-block:: yaml
|
.. code-block:: yaml
|
||||||
|
|
||||||
@@ -22,7 +23,8 @@ you can add it as an entry to ``upstreams.yaml``:
|
|||||||
install_tree: /path/to/another/spack/opt/spack
|
install_tree: /path/to/another/spack/opt/spack
|
||||||
|
|
||||||
``install_tree`` must point to the ``opt/spack`` directory inside of the
|
``install_tree`` must point to the ``opt/spack`` directory inside of the
|
||||||
Spack base directory.
|
Spack base directory, or the location of the ``install_tree`` defined
|
||||||
|
in :ref:`config.yaml <config-yaml>`.
|
||||||
|
|
||||||
Once the upstream Spack instance has been added, ``spack find`` will
|
Once the upstream Spack instance has been added, ``spack find`` will
|
||||||
automatically check the upstream instance when querying installed packages,
|
automatically check the upstream instance when querying installed packages,
|
||||||
|
|||||||
@@ -203,12 +203,9 @@ The OS that are currently supported are summarized in the table below:
|
|||||||
* - Ubuntu 24.04
|
* - Ubuntu 24.04
|
||||||
- ``ubuntu:24.04``
|
- ``ubuntu:24.04``
|
||||||
- ``spack/ubuntu-noble``
|
- ``spack/ubuntu-noble``
|
||||||
* - CentOS 7
|
* - CentOS Stream9
|
||||||
- ``centos:7``
|
- ``quay.io/centos/centos:stream9``
|
||||||
- ``spack/centos7``
|
- ``spack/centos-stream9``
|
||||||
* - CentOS Stream
|
|
||||||
- ``quay.io/centos/centos:stream``
|
|
||||||
- ``spack/centos-stream``
|
|
||||||
* - openSUSE Leap
|
* - openSUSE Leap
|
||||||
- ``opensuse/leap``
|
- ``opensuse/leap``
|
||||||
- ``spack/leap15``
|
- ``spack/leap15``
|
||||||
|
|||||||
@@ -460,6 +460,125 @@ Sourcing that file in Bash will make the environment available to the
|
|||||||
user; and can be included in ``.bashrc`` files, etc. The ``loads``
|
user; and can be included in ``.bashrc`` files, etc. The ``loads``
|
||||||
file may also be copied out of the environment, renamed, etc.
|
file may also be copied out of the environment, renamed, etc.
|
||||||
|
|
||||||
|
|
||||||
|
.. _environment_include_concrete:
|
||||||
|
|
||||||
|
------------------------------
|
||||||
|
Included Concrete Environments
|
||||||
|
------------------------------
|
||||||
|
|
||||||
|
Spack environments can create an environment based off of information in already
|
||||||
|
established environments. You can think of it as a combination of existing
|
||||||
|
environments. It will gather information from the existing environment's
|
||||||
|
``spack.lock`` and use that during the creation of this included concrete
|
||||||
|
environment. When an included concrete environment is created it will generate
|
||||||
|
a ``spack.lock`` file for the newly created environment.
|
||||||
|
|
||||||
|
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
Creating included environments
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
To create a combined concrete environment, you must have at least one existing
|
||||||
|
concrete environment. You will use the command ``spack env create`` with the
|
||||||
|
argument ``--include-concrete`` followed by the name or path of the environment
|
||||||
|
you'd like to include. Here is an example of how to create a combined environment
|
||||||
|
from the command line.
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
$ spack env create myenv
|
||||||
|
$ spack -e myenv add python
|
||||||
|
$ spack -e myenv concretize
|
||||||
|
$ spack env create --include-concrete myenv included_env
|
||||||
|
|
||||||
|
|
||||||
|
You can also include an environment directly in the ``spack.yaml`` file. It
|
||||||
|
involves adding the ``include_concrete`` heading in the yaml followed by the
|
||||||
|
absolute path to the independent environments.
|
||||||
|
|
||||||
|
.. code-block:: yaml
|
||||||
|
|
||||||
|
spack:
|
||||||
|
specs: []
|
||||||
|
concretizer:
|
||||||
|
unify: true
|
||||||
|
include_concrete:
|
||||||
|
- /absolute/path/to/environment1
|
||||||
|
- /absolute/path/to/environment2
|
||||||
|
|
||||||
|
|
||||||
|
Once the ``spack.yaml`` has been updated you must concretize the environment to
|
||||||
|
get the concrete specs from the included environments.
|
||||||
|
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
Updating an included environment
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
If changes were made to the base environment and you want that reflected in the
|
||||||
|
included environment you will need to reconcretize both the base environment and the
|
||||||
|
included environment for the change to be implemented. For example:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
$ spack env create myenv
|
||||||
|
$ spack -e myenv add python
|
||||||
|
$ spack -e myenv concretize
|
||||||
|
$ spack env create --include-concrete myenv included_env
|
||||||
|
|
||||||
|
|
||||||
|
$ spack -e myenv find
|
||||||
|
==> In environment myenv
|
||||||
|
==> Root specs
|
||||||
|
python
|
||||||
|
|
||||||
|
==> 0 installed packages
|
||||||
|
|
||||||
|
|
||||||
|
$ spack -e included_env find
|
||||||
|
==> In environment included_env
|
||||||
|
==> No root specs
|
||||||
|
==> Included specs
|
||||||
|
python
|
||||||
|
|
||||||
|
==> 0 installed packages
|
||||||
|
|
||||||
|
Here we see that ``included_env`` has access to the python package through
|
||||||
|
the ``myenv`` environment. But if we were to add another spec to ``myenv``,
|
||||||
|
``included_env`` will not be able to access the new information.
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
$ spack -e myenv add perl
|
||||||
|
$ spack -e myenv concretize
|
||||||
|
$ spack -e myenv find
|
||||||
|
==> In environment myenv
|
||||||
|
==> Root specs
|
||||||
|
perl python
|
||||||
|
|
||||||
|
==> 0 installed packages
|
||||||
|
|
||||||
|
|
||||||
|
$ spack -e included_env find
|
||||||
|
==> In environment included_env
|
||||||
|
==> No root specs
|
||||||
|
==> Included specs
|
||||||
|
python
|
||||||
|
|
||||||
|
==> 0 installed packages
|
||||||
|
|
||||||
|
It isn't until you run the ``spack concretize`` command that the combined
|
||||||
|
environment will get the updated information from the reconcretized base environmennt.
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
$ spack -e included_env concretize
|
||||||
|
$ spack -e included_env find
|
||||||
|
==> In environment included_env
|
||||||
|
==> No root specs
|
||||||
|
==> Included specs
|
||||||
|
perl python
|
||||||
|
|
||||||
|
==> 0 installed packages
|
||||||
|
|
||||||
.. _environment-configuration:
|
.. _environment-configuration:
|
||||||
|
|
||||||
------------------------
|
------------------------
|
||||||
@@ -811,32 +930,85 @@ For example, the following environment has three root packages:
|
|||||||
This allows for a much-needed reduction in redundancy between packages
|
This allows for a much-needed reduction in redundancy between packages
|
||||||
and constraints.
|
and constraints.
|
||||||
|
|
||||||
----------------
|
|
||||||
Filesystem Views
|
|
||||||
----------------
|
|
||||||
|
|
||||||
Spack Environments can define filesystem views, which provide a direct access point
|
-----------------
|
||||||
for software similar to the directory hierarchy that might exist under ``/usr/local``.
|
Environment Views
|
||||||
Filesystem views are updated every time the environment is written out to the lock
|
-----------------
|
||||||
file ``spack.lock``, so the concrete environment and the view are always compatible.
|
|
||||||
The files of the view's installed packages are brought into the view by symbolic or
|
Spack Environments can have an associated filesystem view, which is a directory
|
||||||
hard links, referencing the original Spack installation, or by copy.
|
with a more traditional structure ``<view>/bin``, ``<view>/lib``, ``<view>/include``
|
||||||
|
in which all files of the installed packages are linked.
|
||||||
|
|
||||||
|
By default a view is created for each environment, thanks to the ``view: true``
|
||||||
|
option in the ``spack.yaml`` manifest file:
|
||||||
|
|
||||||
|
.. code-block:: yaml
|
||||||
|
|
||||||
|
spack:
|
||||||
|
specs: [perl, python]
|
||||||
|
view: true
|
||||||
|
|
||||||
|
The view is created in a hidden directory ``.spack-env/view`` relative to the environment.
|
||||||
|
If you've used ``spack env activate``, you may have already interacted with this view. Spack
|
||||||
|
prepends its ``<view>/bin`` dir to ``PATH`` when the environment is activated, so that
|
||||||
|
you can directly run executables from all installed packages in the environment.
|
||||||
|
|
||||||
|
Views are highly customizable: you can control where they are put, modify their structure,
|
||||||
|
include and exclude specs, change how files are linked, and you can even generate multiple
|
||||||
|
views for a single environment.
|
||||||
|
|
||||||
.. _configuring_environment_views:
|
.. _configuring_environment_views:
|
||||||
|
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
Configuration in ``spack.yaml``
|
Minimal view configuration
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
The Spack Environment manifest file has a top-level keyword
|
The minimal configuration
|
||||||
``view``. Each entry under that heading is a **view descriptor**, headed
|
|
||||||
by a name. Any number of views may be defined under the ``view`` heading.
|
.. code-block:: yaml
|
||||||
The view descriptor contains the root of the view, and
|
|
||||||
optionally the projections for the view, ``select`` and
|
spack:
|
||||||
``exclude`` lists for the view and link information via ``link`` and
|
# ...
|
||||||
|
view: true
|
||||||
|
|
||||||
|
lets Spack generate a single view with default settings under the
|
||||||
|
``.spack-env/view`` directory of the environment.
|
||||||
|
|
||||||
|
Another short way to configure a view is to specify just where to put it:
|
||||||
|
|
||||||
|
.. code-block:: yaml
|
||||||
|
|
||||||
|
spack:
|
||||||
|
# ...
|
||||||
|
view: /path/to/view
|
||||||
|
|
||||||
|
Views can also be disabled by setting ``view: false``.
|
||||||
|
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
Advanced view configuration
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
|
One or more **view descriptors** can be defined under ``view``, keyed by a name.
|
||||||
|
The example from the previous section with ``view: /path/to/view`` is equivalent
|
||||||
|
to defining a view descriptor named ``default`` with a ``root`` attribute:
|
||||||
|
|
||||||
|
.. code-block:: yaml
|
||||||
|
|
||||||
|
spack:
|
||||||
|
# ...
|
||||||
|
view:
|
||||||
|
default: # name of the view
|
||||||
|
root: /path/to/view # view descriptor attribute
|
||||||
|
|
||||||
|
The ``default`` view descriptor name is special: when you ``spack env activate`` your
|
||||||
|
environment, this view will be used to update (among other things) your ``PATH``
|
||||||
|
variable.
|
||||||
|
|
||||||
|
View descriptors must contain the root of the view, and optionally projections,
|
||||||
|
``select`` and ``exclude`` lists and link information via ``link`` and
|
||||||
``link_type``.
|
``link_type``.
|
||||||
|
|
||||||
For example, in the following manifest
|
As a more advanced example, in the following manifest
|
||||||
file snippet we define a view named ``mpis``, rooted at
|
file snippet we define a view named ``mpis``, rooted at
|
||||||
``/path/to/view`` in which all projections use the package name,
|
``/path/to/view`` in which all projections use the package name,
|
||||||
version, and compiler name to determine the path for a given
|
version, and compiler name to determine the path for a given
|
||||||
@@ -881,59 +1053,10 @@ of ``hardlink`` or ``copy``.
|
|||||||
when the environment is not activated, and linked libraries will be located
|
when the environment is not activated, and linked libraries will be located
|
||||||
*outside* of the view thanks to rpaths.
|
*outside* of the view thanks to rpaths.
|
||||||
|
|
||||||
|
|
||||||
There are two shorthands for environments with a single view. If the
|
|
||||||
environment at ``/path/to/env`` has a single view, with a root at
|
|
||||||
``/path/to/env/.spack-env/view``, with default selection and exclusion
|
|
||||||
and the default projection, we can put ``view: True`` in the
|
|
||||||
environment manifest. Similarly, if the environment has a view with a
|
|
||||||
different root, but default selection, exclusion, and projections, the
|
|
||||||
manifest can say ``view: /path/to/view``. These views are
|
|
||||||
automatically named ``default``, so that
|
|
||||||
|
|
||||||
.. code-block:: yaml
|
|
||||||
|
|
||||||
spack:
|
|
||||||
# ...
|
|
||||||
view: True
|
|
||||||
|
|
||||||
is equivalent to
|
|
||||||
|
|
||||||
.. code-block:: yaml
|
|
||||||
|
|
||||||
spack:
|
|
||||||
# ...
|
|
||||||
view:
|
|
||||||
default:
|
|
||||||
root: .spack-env/view
|
|
||||||
|
|
||||||
and
|
|
||||||
|
|
||||||
.. code-block:: yaml
|
|
||||||
|
|
||||||
spack:
|
|
||||||
# ...
|
|
||||||
view: /path/to/view
|
|
||||||
|
|
||||||
is equivalent to
|
|
||||||
|
|
||||||
.. code-block:: yaml
|
|
||||||
|
|
||||||
spack:
|
|
||||||
# ...
|
|
||||||
view:
|
|
||||||
default:
|
|
||||||
root: /path/to/view
|
|
||||||
|
|
||||||
By default, Spack environments are configured with ``view: True`` in
|
|
||||||
the manifest. Environments can be configured without views using
|
|
||||||
``view: False``. For backwards compatibility reasons, environments
|
|
||||||
with no ``view`` key are treated the same as ``view: True``.
|
|
||||||
|
|
||||||
From the command line, the ``spack env create`` command takes an
|
From the command line, the ``spack env create`` command takes an
|
||||||
argument ``--with-view [PATH]`` that sets the path for a single, default
|
argument ``--with-view [PATH]`` that sets the path for a single, default
|
||||||
view. If no path is specified, the default path is used (``view:
|
view. If no path is specified, the default path is used (``view:
|
||||||
True``). The argument ``--without-view`` can be used to create an
|
true``). The argument ``--without-view`` can be used to create an
|
||||||
environment without any view configured.
|
environment without any view configured.
|
||||||
|
|
||||||
The ``spack env view`` command can be used to change the manage views
|
The ``spack env view`` command can be used to change the manage views
|
||||||
@@ -999,11 +1122,18 @@ the projection under ``all`` before reaching those entries.
|
|||||||
Activating environment views
|
Activating environment views
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
The ``spack env activate`` command will put the default view for the
|
The ``spack env activate <env>`` has two effects:
|
||||||
environment into the user's path, in addition to activating the
|
|
||||||
environment for Spack commands. The arguments ``-v,--with-view`` and
|
1. It activates the environment so that further Spack commands such
|
||||||
``-V,--without-view`` can be used to tune this behavior. The default
|
as ``spack install`` will run in the context of the environment.
|
||||||
behavior is to activate with the environment view if there is one.
|
2. It activates the view so that environment variables such as
|
||||||
|
``PATH`` are updated to include the view.
|
||||||
|
|
||||||
|
Without further arguments, the ``default`` view of the environment is
|
||||||
|
activated. If a view with a different name has to be activated,
|
||||||
|
``spack env activate --with-view <name> <env>`` can be
|
||||||
|
used instead. You can also activate the environment without modifying
|
||||||
|
further environment variables using ``--without-view``.
|
||||||
|
|
||||||
The environment variables affected by the ``spack env activate``
|
The environment variables affected by the ``spack env activate``
|
||||||
command and the paths that are used to update them are determined by
|
command and the paths that are used to update them are determined by
|
||||||
@@ -1026,8 +1156,8 @@ relevant variable if the path exists. For this reason, it is not
|
|||||||
recommended to use non-default projections with the default view of an
|
recommended to use non-default projections with the default view of an
|
||||||
environment.
|
environment.
|
||||||
|
|
||||||
The ``spack env deactivate`` command will remove the default view of
|
The ``spack env deactivate`` command will remove the active view of
|
||||||
the environment from the user's path.
|
the Spack environment from the user's environment variables.
|
||||||
|
|
||||||
|
|
||||||
.. _env-generate-depfile:
|
.. _env-generate-depfile:
|
||||||
@@ -1044,7 +1174,7 @@ other targets to depend on the environment installation.
|
|||||||
|
|
||||||
A typical workflow is as follows:
|
A typical workflow is as follows:
|
||||||
|
|
||||||
.. code:: console
|
.. code-block:: console
|
||||||
|
|
||||||
spack env create -d .
|
spack env create -d .
|
||||||
spack -e . add perl
|
spack -e . add perl
|
||||||
@@ -1137,7 +1267,7 @@ its dependencies. This can be useful when certain flags should only apply to
|
|||||||
dependencies. Below we show a use case where a spec is installed with verbose
|
dependencies. Below we show a use case where a spec is installed with verbose
|
||||||
output (``spack install --verbose``) while its dependencies are installed silently:
|
output (``spack install --verbose``) while its dependencies are installed silently:
|
||||||
|
|
||||||
.. code:: console
|
.. code-block:: console
|
||||||
|
|
||||||
$ spack env depfile -o Makefile
|
$ spack env depfile -o Makefile
|
||||||
|
|
||||||
@@ -1159,7 +1289,7 @@ This can be accomplished through the generated ``[<prefix>/]SPACK_PACKAGE_IDS``
|
|||||||
variable. Assuming we have an active and concrete environment, we generate the
|
variable. Assuming we have an active and concrete environment, we generate the
|
||||||
associated ``Makefile`` with a prefix ``example``:
|
associated ``Makefile`` with a prefix ``example``:
|
||||||
|
|
||||||
.. code:: console
|
.. code-block:: console
|
||||||
|
|
||||||
$ spack env depfile -o env.mk --make-prefix example
|
$ spack env depfile -o env.mk --make-prefix example
|
||||||
|
|
||||||
@@ -1186,7 +1316,7 @@ index once every package is pushed. Note how this target uses the generated
|
|||||||
example/push/%: example/install/%
|
example/push/%: example/install/%
|
||||||
@mkdir -p $(dir $@)
|
@mkdir -p $(dir $@)
|
||||||
$(info About to push $(SPEC) to a buildcache)
|
$(info About to push $(SPEC) to a buildcache)
|
||||||
$(SPACK) -e . buildcache push --allow-root --only=package $(BUILDCACHE_DIR) /$(HASH)
|
$(SPACK) -e . buildcache push --only=package $(BUILDCACHE_DIR) /$(HASH)
|
||||||
@touch $@
|
@touch $@
|
||||||
|
|
||||||
push: $(addprefix example/push/,$(example/SPACK_PACKAGE_IDS))
|
push: $(addprefix example/push/,$(example/SPACK_PACKAGE_IDS))
|
||||||
|
|||||||
@@ -1364,187 +1364,6 @@ This will write the private key to the file `dinosaur.priv`.
|
|||||||
or for help on an issue or the Spack slack.
|
or for help on an issue or the Spack slack.
|
||||||
|
|
||||||
|
|
||||||
.. _cray-support:
|
|
||||||
|
|
||||||
-------------
|
|
||||||
Spack on Cray
|
|
||||||
-------------
|
|
||||||
|
|
||||||
Spack differs slightly when used on a Cray system. The architecture spec
|
|
||||||
can differentiate between the front-end and back-end processor and operating system.
|
|
||||||
For example, on Edison at NERSC, the back-end target processor
|
|
||||||
is "Ivy Bridge", so you can specify to use the back-end this way:
|
|
||||||
|
|
||||||
.. code-block:: console
|
|
||||||
|
|
||||||
$ spack install zlib target=ivybridge
|
|
||||||
|
|
||||||
You can also use the operating system to build against the back-end:
|
|
||||||
|
|
||||||
.. code-block:: console
|
|
||||||
|
|
||||||
$ spack install zlib os=CNL10
|
|
||||||
|
|
||||||
Notice that the name includes both the operating system name and the major
|
|
||||||
version number concatenated together.
|
|
||||||
|
|
||||||
Alternatively, if you want to build something for the front-end,
|
|
||||||
you can specify the front-end target processor. The processor for a login node
|
|
||||||
on Edison is "Sandy bridge" so we specify on the command line like so:
|
|
||||||
|
|
||||||
.. code-block:: console
|
|
||||||
|
|
||||||
$ spack install zlib target=sandybridge
|
|
||||||
|
|
||||||
And the front-end operating system is:
|
|
||||||
|
|
||||||
.. code-block:: console
|
|
||||||
|
|
||||||
$ spack install zlib os=SuSE11
|
|
||||||
|
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^
|
|
||||||
Cray compiler detection
|
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^
|
|
||||||
|
|
||||||
Spack can detect compilers using two methods. For the front-end, we treat
|
|
||||||
everything the same. The difference lies in back-end compiler detection.
|
|
||||||
Back-end compiler detection is made via the Tcl module avail command.
|
|
||||||
Once it detects the compiler it writes the appropriate PrgEnv and compiler
|
|
||||||
module name to compilers.yaml and sets the paths to each compiler with Cray\'s
|
|
||||||
compiler wrapper names (i.e. cc, CC, ftn). During build time, Spack will load
|
|
||||||
the correct PrgEnv and compiler module and will call appropriate wrapper.
|
|
||||||
|
|
||||||
The compilers.yaml config file will also differ. There is a
|
|
||||||
modules section that is filled with the compiler's Programming Environment
|
|
||||||
and module name. On other systems, this field is empty []:
|
|
||||||
|
|
||||||
.. code-block:: yaml
|
|
||||||
|
|
||||||
- compiler:
|
|
||||||
modules:
|
|
||||||
- PrgEnv-intel
|
|
||||||
- intel/15.0.109
|
|
||||||
|
|
||||||
As mentioned earlier, the compiler paths will look different on a Cray system.
|
|
||||||
Since most compilers are invoked using cc, CC and ftn, the paths for each
|
|
||||||
compiler are replaced with their respective Cray compiler wrapper names:
|
|
||||||
|
|
||||||
.. code-block:: yaml
|
|
||||||
|
|
||||||
paths:
|
|
||||||
cc: cc
|
|
||||||
cxx: CC
|
|
||||||
f77: ftn
|
|
||||||
fc: ftn
|
|
||||||
|
|
||||||
As opposed to an explicit path to the compiler executable. This allows Spack
|
|
||||||
to call the Cray compiler wrappers during build time.
|
|
||||||
|
|
||||||
For more on compiler configuration, check out :ref:`compiler-config`.
|
|
||||||
|
|
||||||
Spack sets the default Cray link type to dynamic, to better match other
|
|
||||||
other platforms. Individual packages can enable static linking (which is the
|
|
||||||
default outside of Spack on cray systems) using the ``-static`` flag.
|
|
||||||
|
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
|
||||||
Setting defaults and using Cray modules
|
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
|
||||||
|
|
||||||
If you want to use default compilers for each PrgEnv and also be able
|
|
||||||
to load cray external modules, you will need to set up a ``packages.yaml``.
|
|
||||||
|
|
||||||
Here's an example of an external configuration for cray modules:
|
|
||||||
|
|
||||||
.. code-block:: yaml
|
|
||||||
|
|
||||||
packages:
|
|
||||||
mpich:
|
|
||||||
externals:
|
|
||||||
- spec: "mpich@7.3.1%gcc@5.2.0 arch=cray_xc-haswell-CNL10"
|
|
||||||
modules:
|
|
||||||
- cray-mpich
|
|
||||||
- spec: "mpich@7.3.1%intel@16.0.0.109 arch=cray_xc-haswell-CNL10"
|
|
||||||
modules:
|
|
||||||
- cray-mpich
|
|
||||||
all:
|
|
||||||
providers:
|
|
||||||
mpi: [mpich]
|
|
||||||
|
|
||||||
This tells Spack that for whatever package that depends on mpi, load the
|
|
||||||
cray-mpich module into the environment. You can then be able to use whatever
|
|
||||||
environment variables, libraries, etc, that are brought into the environment
|
|
||||||
via module load.
|
|
||||||
|
|
||||||
.. note::
|
|
||||||
|
|
||||||
For Cray-provided packages, it is best to use ``modules:`` instead of ``prefix:``
|
|
||||||
in ``packages.yaml``, because the Cray Programming Environment heavily relies on
|
|
||||||
modules (e.g., loading the ``cray-mpich`` module adds MPI libraries to the
|
|
||||||
compiler wrapper link line).
|
|
||||||
|
|
||||||
You can set the default compiler that Spack can use for each compiler type.
|
|
||||||
If you want to use the Cray defaults, then set them under ``all:`` in packages.yaml.
|
|
||||||
In the compiler field, set the compiler specs in your order of preference.
|
|
||||||
Whenever you build with that compiler type, Spack will concretize to that version.
|
|
||||||
|
|
||||||
Here is an example of a full packages.yaml used at NERSC
|
|
||||||
|
|
||||||
.. code-block:: yaml
|
|
||||||
|
|
||||||
packages:
|
|
||||||
mpich:
|
|
||||||
externals:
|
|
||||||
- spec: "mpich@7.3.1%gcc@5.2.0 arch=cray_xc-CNL10-ivybridge"
|
|
||||||
modules:
|
|
||||||
- cray-mpich
|
|
||||||
- spec: "mpich@7.3.1%intel@16.0.0.109 arch=cray_xc-SuSE11-ivybridge"
|
|
||||||
modules:
|
|
||||||
- cray-mpich
|
|
||||||
buildable: False
|
|
||||||
netcdf:
|
|
||||||
externals:
|
|
||||||
- spec: "netcdf@4.3.3.1%gcc@5.2.0 arch=cray_xc-CNL10-ivybridge"
|
|
||||||
modules:
|
|
||||||
- cray-netcdf
|
|
||||||
- spec: "netcdf@4.3.3.1%intel@16.0.0.109 arch=cray_xc-CNL10-ivybridge"
|
|
||||||
modules:
|
|
||||||
- cray-netcdf
|
|
||||||
buildable: False
|
|
||||||
hdf5:
|
|
||||||
externals:
|
|
||||||
- spec: "hdf5@1.8.14%gcc@5.2.0 arch=cray_xc-CNL10-ivybridge"
|
|
||||||
modules:
|
|
||||||
- cray-hdf5
|
|
||||||
- spec: "hdf5@1.8.14%intel@16.0.0.109 arch=cray_xc-CNL10-ivybridge"
|
|
||||||
modules:
|
|
||||||
- cray-hdf5
|
|
||||||
buildable: False
|
|
||||||
all:
|
|
||||||
compiler: [gcc@5.2.0, intel@16.0.0.109]
|
|
||||||
providers:
|
|
||||||
mpi: [mpich]
|
|
||||||
|
|
||||||
Here we tell spack that whenever we want to build with gcc use version 5.2.0 or
|
|
||||||
if we want to build with intel compilers, use version 16.0.0.109. We add a spec
|
|
||||||
for each compiler type for each cray modules. This ensures that for each
|
|
||||||
compiler on our system we can use that external module.
|
|
||||||
|
|
||||||
For more on external packages check out the section :ref:`sec-external-packages`.
|
|
||||||
|
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
|
||||||
Using Linux containers on Cray machines
|
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
|
||||||
|
|
||||||
Spack uses environment variables particular to the Cray programming
|
|
||||||
environment to determine which systems are Cray platforms. These
|
|
||||||
environment variables may be propagated into containers that are not
|
|
||||||
using the Cray programming environment.
|
|
||||||
|
|
||||||
To ensure that Spack does not autodetect the Cray programming
|
|
||||||
environment, unset the environment variable ``MODULEPATH``. This
|
|
||||||
will cause Spack to treat a linux container on a Cray system as a base
|
|
||||||
linux distro.
|
|
||||||
|
|
||||||
.. _windows_support:
|
.. _windows_support:
|
||||||
|
|
||||||
----------------
|
----------------
|
||||||
|
|||||||
@@ -2344,6 +2344,27 @@ you set ``parallel`` to ``False`` at the package level, then each call
|
|||||||
to ``make()`` will be sequential by default, but packagers can call
|
to ``make()`` will be sequential by default, but packagers can call
|
||||||
``make(parallel=True)`` to override it.
|
``make(parallel=True)`` to override it.
|
||||||
|
|
||||||
|
Note that the ``--jobs`` option works out of the box for all standard
|
||||||
|
build systems. If you are using a non-standard build system instead, you
|
||||||
|
can use the variable ``make_jobs`` to extract the number of jobs specified
|
||||||
|
by the ``--jobs`` option:
|
||||||
|
|
||||||
|
.. code-block:: python
|
||||||
|
:emphasize-lines: 7, 11
|
||||||
|
:linenos:
|
||||||
|
|
||||||
|
class Xios(Package):
|
||||||
|
...
|
||||||
|
def install(self, spec, prefix):
|
||||||
|
...
|
||||||
|
options = [
|
||||||
|
...
|
||||||
|
'--jobs', str(make_jobs),
|
||||||
|
]
|
||||||
|
...
|
||||||
|
make_xios = Executable("./make_xios")
|
||||||
|
make_xios(*options)
|
||||||
|
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
Install-level build parallelism
|
Install-level build parallelism
|
||||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
@@ -5173,12 +5194,6 @@ installed executable. The check is implemented as follows:
|
|||||||
reframe = Executable(self.prefix.bin.reframe)
|
reframe = Executable(self.prefix.bin.reframe)
|
||||||
reframe("-l")
|
reframe("-l")
|
||||||
|
|
||||||
.. warning::
|
|
||||||
|
|
||||||
The API for adding tests is not yet considered stable and may change
|
|
||||||
in future releases.
|
|
||||||
|
|
||||||
|
|
||||||
""""""""""""""""""""""""""""""""
|
""""""""""""""""""""""""""""""""
|
||||||
Checking build-time test results
|
Checking build-time test results
|
||||||
""""""""""""""""""""""""""""""""
|
""""""""""""""""""""""""""""""""
|
||||||
@@ -5216,38 +5231,42 @@ be left in the build stage directory as illustrated below:
|
|||||||
Stand-alone tests
|
Stand-alone tests
|
||||||
^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
While build-time tests are integrated with the build process, stand-alone
|
While build-time tests are integrated with the installation process, stand-alone
|
||||||
tests are expected to run days, weeks, even months after the software is
|
tests are expected to run days, weeks, even months after the software is
|
||||||
installed. The goal is to provide a mechanism for gaining confidence that
|
installed. The goal is to provide a mechanism for gaining confidence that
|
||||||
packages work as installed **and** *continue* to work as the underlying
|
packages work as installed **and** *continue* to work as the underlying
|
||||||
software evolves. Packages can add and inherit stand-alone tests. The
|
software evolves. Packages can add and inherit stand-alone tests. The
|
||||||
`spack test`` command is used to manage stand-alone testing.
|
``spack test`` command is used for stand-alone testing.
|
||||||
|
|
||||||
.. note::
|
.. admonition:: Stand-alone test methods should complete within a few minutes.
|
||||||
|
|
||||||
Execution speed is important since these tests are intended to quickly
|
Execution speed is important since these tests are intended to quickly
|
||||||
assess whether installed specs work on the system. Consequently, they
|
assess whether installed specs work on the system. Spack cannot spare
|
||||||
should run relatively quickly -- as in on the order of at most a few
|
resources for more extensive testing of packages included in CI stacks.
|
||||||
minutes -- while ideally executing all, or at least key aspects of the
|
|
||||||
installed software.
|
|
||||||
|
|
||||||
.. note::
|
Consequently, stand-alone tests should run relatively quickly -- as in
|
||||||
|
on the order of at most a few minutes -- while testing at least key aspects
|
||||||
Failing stand-alone tests indicate problems with the installation and,
|
of the installed software. Save more extensive testing for other tools.
|
||||||
therefore, there is no reason to proceed with more resource-intensive
|
|
||||||
tests until those have been investigated.
|
|
||||||
|
|
||||||
Passing stand-alone tests indicate that more thorough testing, such
|
|
||||||
as running extensive unit or regression tests, or tests that run at
|
|
||||||
scale can proceed without wasting resources on a problematic installation.
|
|
||||||
|
|
||||||
Tests are defined in the package using methods with names beginning ``test_``.
|
Tests are defined in the package using methods with names beginning ``test_``.
|
||||||
This allows Spack to support multiple independent checks, or parts. Files
|
This allows Spack to support multiple independent checks, or parts. Files
|
||||||
needed for testing, such as source, data, and expected outputs, may be saved
|
needed for testing, such as source, data, and expected outputs, may be saved
|
||||||
from the build and or stored with the package in the repository. Regardless
|
from the build and or stored with the package in the repository. Regardless
|
||||||
of origin, these files are automatically copied to the spec's test stage
|
of origin, these files are automatically copied to the spec's test stage
|
||||||
directory prior to execution of the test method(s). Spack also provides some
|
directory prior to execution of the test method(s). Spack also provides helper
|
||||||
helper functions to facilitate processing.
|
functions to facilitate common processing.
|
||||||
|
|
||||||
|
.. tip::
|
||||||
|
|
||||||
|
**The status of stand-alone tests can be used to guide follow-up testing efforts.**
|
||||||
|
|
||||||
|
Passing stand-alone tests justify performing more thorough testing, such
|
||||||
|
as running extensive unit or regression tests or tests that run at scale,
|
||||||
|
when available. These tests are outside of the scope of Spack packaging.
|
||||||
|
|
||||||
|
Failing stand-alone tests indicate problems with the installation and,
|
||||||
|
therefore, no reason to proceed with more resource-intensive tests until
|
||||||
|
the failures have been investigated.
|
||||||
|
|
||||||
.. _configure-test-stage:
|
.. _configure-test-stage:
|
||||||
|
|
||||||
@@ -5255,30 +5274,26 @@ helper functions to facilitate processing.
|
|||||||
Configuring the test stage directory
|
Configuring the test stage directory
|
||||||
""""""""""""""""""""""""""""""""""""
|
""""""""""""""""""""""""""""""""""""
|
||||||
|
|
||||||
Stand-alone tests utilize a test stage directory for building, running,
|
Stand-alone tests utilize a test stage directory to build, run, and track
|
||||||
and tracking results in the same way Spack uses a build stage directory.
|
tests in the same way Spack uses a build stage directory to install software.
|
||||||
The default test stage root directory, ``~/.spack/test``, is defined in
|
The default test stage root directory, ``$HOME/.spack/test``, is defined in
|
||||||
:ref:`etc/spack/defaults/config.yaml <config-yaml>`. This location is
|
:ref:`config.yaml <config-yaml>`. This location is customizable by adding or
|
||||||
customizable by adding or changing the ``test_stage`` path in the high-level
|
changing the ``test_stage`` path such that:
|
||||||
``config`` of the appropriate ``config.yaml`` file such that:
|
|
||||||
|
|
||||||
.. code-block:: yaml
|
.. code-block:: yaml
|
||||||
|
|
||||||
config:
|
config:
|
||||||
test_stage: /path/to/test/stage
|
test_stage: /path/to/test/stage
|
||||||
|
|
||||||
Packages can use the ``self.test_suite.stage`` property to access this setting.
|
Packages can use the ``self.test_suite.stage`` property to access the path.
|
||||||
Other package properties that provide access to spec-specific subdirectories
|
|
||||||
and files are described in :ref:`accessing staged files <accessing-files>`.
|
|
||||||
|
|
||||||
.. note::
|
.. admonition:: Each spec being tested has its own test stage directory.
|
||||||
|
|
||||||
The test stage path is the root directory for the **entire suite**.
|
The ``config:test_stage`` option is the path to the root of a
|
||||||
In other words, it is the root directory for **all specs** being
|
**test suite**'s stage directories.
|
||||||
tested by the ``spack test run`` command. Each spec gets its own
|
|
||||||
stage subdirectory. Use ``self.test_suite.test_dir_for_spec(self.spec)``
|
|
||||||
to access the spec-specific test stage directory.
|
|
||||||
|
|
||||||
|
Other package properties that provide paths to spec-specific subdirectories
|
||||||
|
and files are described in :ref:`accessing-files`.
|
||||||
|
|
||||||
.. _adding-standalone-tests:
|
.. _adding-standalone-tests:
|
||||||
|
|
||||||
@@ -5291,61 +5306,144 @@ Test recipes are defined in the package using methods with names beginning
|
|||||||
Each method has access to the information Spack tracks on the package, such
|
Each method has access to the information Spack tracks on the package, such
|
||||||
as options, compilers, and dependencies, supporting the customization of tests
|
as options, compilers, and dependencies, supporting the customization of tests
|
||||||
to the build. Standard python ``assert`` statements and other error reporting
|
to the build. Standard python ``assert`` statements and other error reporting
|
||||||
mechanisms are available. Such exceptions are automatically caught and reported
|
mechanisms can be used. These exceptions are automatically caught and reported
|
||||||
as test failures.
|
as test failures.
|
||||||
|
|
||||||
Each test method is an implicit test part named by the method and whose
|
Each test method is an *implicit test part* named by the method. Its purpose
|
||||||
purpose is the method's docstring. Providing a purpose gives context for
|
is the method's docstring. Providing a meaningful purpose for the test gives
|
||||||
aiding debugging. A test method may contain embedded test parts. Spack
|
context that can aid debugging. Spack outputs both the name and purpose at the
|
||||||
outputs the test name and purpose prior to running each test method and
|
start of test execution so it's also important that the docstring/purpose be
|
||||||
any embedded test parts. For example, ``MyPackage`` below provides two basic
|
brief.
|
||||||
examples of installation tests: ``test_always_fails`` and ``test_example``.
|
|
||||||
As the name indicates, the first always fails. The second simply runs the
|
.. tip::
|
||||||
installed example.
|
|
||||||
|
We recommend naming test methods so it is clear *what* is being tested.
|
||||||
|
For example, if a test method is building and or running an executable
|
||||||
|
called ``example``, then call the method ``test_example``. This, together
|
||||||
|
with a similarly meaningful test purpose, will aid test comprehension,
|
||||||
|
debugging, and maintainability.
|
||||||
|
|
||||||
|
Stand-alone tests run in an environment that provides access to information
|
||||||
|
on the installed software, such as build options, dependencies, and compilers.
|
||||||
|
Build options and dependencies are accessed using the same spec checks used
|
||||||
|
by build recipes. Examples of checking :ref:`variant settings <variants>` and
|
||||||
|
:ref:`spec constraints <testing-specs>` can be found at the provided links.
|
||||||
|
|
||||||
|
.. admonition:: Spack automatically sets up the test stage directory and environment.
|
||||||
|
|
||||||
|
Spack automatically creates the test stage directory and copies
|
||||||
|
relevant files *prior to* running tests. It can also ensure build
|
||||||
|
dependencies are available **if** necessary.
|
||||||
|
|
||||||
|
The path to the test stage is configurable (see :ref:`configure-test-stage`).
|
||||||
|
|
||||||
|
Files that Spack knows to copy are those saved from the build (see
|
||||||
|
:ref:`cache_extra_test_sources`) and those added to the package repository
|
||||||
|
(see :ref:`cache_custom_files`).
|
||||||
|
|
||||||
|
Spack will use the value of the ``test_requires_compiler`` property to
|
||||||
|
determine whether it needs to also set up build dependencies (see
|
||||||
|
:ref:`test-build-tests`).
|
||||||
|
|
||||||
|
The ``MyPackage`` package below provides two basic test examples:
|
||||||
|
``test_example`` and ``test_example2``. The first runs the installed
|
||||||
|
``example`` and ensures its output contains an expected string. The second
|
||||||
|
runs ``example2`` without checking output so is only concerned with confirming
|
||||||
|
the executable runs successfully. If the installed spec is not expected to have
|
||||||
|
``example2``, then the check at the top of the method will raise a special
|
||||||
|
``SkipTest`` exception, which is captured to facilitate reporting skipped test
|
||||||
|
parts to tools like CDash.
|
||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
class MyPackage(Package):
|
class MyPackage(Package):
|
||||||
...
|
...
|
||||||
|
|
||||||
def test_always_fails(self):
|
|
||||||
"""use assert to always fail"""
|
|
||||||
assert False
|
|
||||||
|
|
||||||
def test_example(self):
|
def test_example(self):
|
||||||
"""run installed example"""
|
"""ensure installed example works"""
|
||||||
|
expected = "Done."
|
||||||
example = which(self.prefix.bin.example)
|
example = which(self.prefix.bin.example)
|
||||||
example()
|
|
||||||
|
# Capture stdout and stderr from running the Executable
|
||||||
|
# and check that the expected output was produced.
|
||||||
|
out = example(output=str.split, error=str.split)
|
||||||
|
assert expected in out, f"Expected '{expected}' in the output"
|
||||||
|
|
||||||
|
def test_example2(self):
|
||||||
|
"""run installed example2"""
|
||||||
|
if self.spec.satisfies("@:1.0"):
|
||||||
|
# Raise SkipTest to ensure flagging the test as skipped for
|
||||||
|
# test reporting purposes.
|
||||||
|
raise SkipTest("Test is only available for v1.1 on")
|
||||||
|
|
||||||
|
example2 = which(self.prefix.bin.example2)
|
||||||
|
example2()
|
||||||
|
|
||||||
Output showing the identification of each test part after running the tests
|
Output showing the identification of each test part after running the tests
|
||||||
is illustrated below.
|
is illustrated below.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
$ spack test run --alias mypackage mypackage@1.0
|
$ spack test run --alias mypackage mypackage@2.0
|
||||||
==> Spack test mypackage
|
==> Spack test mypackage
|
||||||
...
|
...
|
||||||
$ spack test results -l mypackage
|
$ spack test results -l mypackage
|
||||||
==> Results for test suite 'mypackage':
|
==> Results for test suite 'mypackage':
|
||||||
...
|
...
|
||||||
==> [2023-03-10-16:03:56.625204] test: test_always_fails: use assert to always fail
|
==> [2024-03-10-16:03:56.625439] test: test_example: ensure installed example works
|
||||||
...
|
...
|
||||||
FAILED
|
PASSED: MyPackage::test_example
|
||||||
==> [2023-03-10-16:03:56.625439] test: test_example: run installed example
|
==> [2024-03-10-16:03:56.625439] test: test_example2: run installed example2
|
||||||
...
|
...
|
||||||
PASSED
|
PASSED: MyPackage::test_example2
|
||||||
|
|
||||||
|
.. admonition:: Do NOT implement tests that must run in the installation prefix.
|
||||||
|
|
||||||
.. note::
|
Use of the package spec's installation prefix for building and running
|
||||||
|
tests is **strongly discouraged**. Doing so causes permission errors for
|
||||||
|
shared spack instances *and* facilities that install the software in
|
||||||
|
read-only file systems or directories.
|
||||||
|
|
||||||
If ``MyPackage`` were a recipe for a library, the tests should build
|
Instead, start these test methods by explicitly copying the needed files
|
||||||
an example or test program that is then executed.
|
from the installation prefix to the test stage directory. Note the test
|
||||||
|
stage directory is the current directory when the test is executed with
|
||||||
|
the ``spack test run`` command.
|
||||||
|
|
||||||
A test method can include test parts using the ``test_part`` context manager.
|
.. admonition:: Test methods for library packages should build test executables.
|
||||||
Each part is treated as an independent check to allow subsequent test parts
|
|
||||||
to execute even after a test part fails.
|
|
||||||
|
|
||||||
.. _test-part:
|
Stand-alone tests for library packages *should* build test executables
|
||||||
|
that utilize the *installed* library. Doing so ensures the tests follow
|
||||||
|
a similar build process that users of the library would follow.
|
||||||
|
|
||||||
|
For more information on how to do this, see :ref:`test-build-tests`.
|
||||||
|
|
||||||
|
.. tip::
|
||||||
|
|
||||||
|
If you want to see more examples from packages with stand-alone tests, run
|
||||||
|
``spack pkg grep "def\stest" | sed "s/\/package.py.*//g" | sort -u``
|
||||||
|
from the command line to get a list of the packages.
|
||||||
|
|
||||||
|
.. _adding-standalone-test-parts:
|
||||||
|
|
||||||
|
"""""""""""""""""""""""""""""
|
||||||
|
Adding stand-alone test parts
|
||||||
|
"""""""""""""""""""""""""""""
|
||||||
|
|
||||||
|
Sometimes dependencies between steps of a test lend themselves to being
|
||||||
|
broken into parts. Tracking the pass/fail status of each part may aid
|
||||||
|
debugging. Spack provides a ``test_part`` context manager for use within
|
||||||
|
test methods.
|
||||||
|
|
||||||
|
Each test part is independently run, tracked, and reported. Test parts are
|
||||||
|
executed in the order they appear. If one fails, subsequent test parts are
|
||||||
|
still performed even if they would also fail. This allows tools like CDash
|
||||||
|
to track and report the status of test parts across runs. The pass/fail status
|
||||||
|
of the enclosing test is derived from the statuses of the embedded test parts.
|
||||||
|
|
||||||
|
.. admonition:: Test method and test part names **must** be unique.
|
||||||
|
|
||||||
|
Test results reporting requires that test methods and embedded test parts
|
||||||
|
within a package have unique names.
|
||||||
|
|
||||||
The signature for ``test_part`` is:
|
The signature for ``test_part`` is:
|
||||||
|
|
||||||
@@ -5367,40 +5465,68 @@ where each argument has the following meaning:
|
|||||||
* ``work_dir`` is the path to the directory in which the test will run.
|
* ``work_dir`` is the path to the directory in which the test will run.
|
||||||
|
|
||||||
The default of ``None``, or ``"."``, corresponds to the the spec's test
|
The default of ``None``, or ``"."``, corresponds to the the spec's test
|
||||||
stage (i.e., ``self.test_suite.test_dir_for_spec(self.spec)``.
|
stage (i.e., ``self.test_suite.test_dir_for_spec(self.spec)``).
|
||||||
|
|
||||||
.. admonition:: Tests should **not** run under the installation directory.
|
.. admonition:: Start test part names with the name of the enclosing test.
|
||||||
|
|
||||||
Use of the package spec's installation directory for building and running
|
We **highly recommend** starting the names of test parts with the name
|
||||||
tests is **strongly** discouraged. Doing so causes permission errors for
|
of the enclosing test. Doing so helps with the comprehension, readability
|
||||||
shared spack instances *and* facilities that install the software in
|
and debugging of test results.
|
||||||
read-only file systems or directories.
|
|
||||||
|
|
||||||
Suppose ``MyPackage`` actually installs two examples we want to use for tests.
|
Suppose ``MyPackage`` installs multiple executables that need to run in a
|
||||||
These checks can be implemented as separate checks or, as illustrated below,
|
specific order since the outputs from one are inputs of others. Further suppose
|
||||||
embedded test parts.
|
we want to add an integration test that runs the executables in order. We can
|
||||||
|
accomplish this goal by implementing a stand-alone test method consisting of
|
||||||
|
test parts for each executable as follows:
|
||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
class MyPackage(Package):
|
class MyPackage(Package):
|
||||||
...
|
...
|
||||||
|
|
||||||
def test_example(self):
|
def test_series(self):
|
||||||
"""run installed examples"""
|
"""run setup, perform, and report"""
|
||||||
for example in ["ex1", "ex2"]:
|
|
||||||
with test_part(
|
|
||||||
self,
|
|
||||||
f"test_example_{example}",
|
|
||||||
purpose=f"run installed {example}",
|
|
||||||
):
|
|
||||||
exe = which(join_path(self.prefix.bin, example))
|
|
||||||
exe()
|
|
||||||
|
|
||||||
In this case, there will be an implicit test part for ``test_example``
|
with test_part(self, "test_series_setup", purpose="setup operation"):
|
||||||
and separate sub-parts for ``ex1`` and ``ex2``. The second sub-part
|
exe = which(self.prefix.bin.setup))
|
||||||
will be executed regardless of whether the first passes. The test
|
exe()
|
||||||
log for a run where the first executable fails and the second passes
|
|
||||||
is illustrated below.
|
with test_part(self, "test_series_run", purpose="perform operation"):
|
||||||
|
exe = which(self.prefix.bin.run))
|
||||||
|
exe()
|
||||||
|
|
||||||
|
with test_part(self, "test_series_report", purpose="generate report"):
|
||||||
|
exe = which(self.prefix.bin.report))
|
||||||
|
exe()
|
||||||
|
|
||||||
|
The result is ``test_series`` runs the following executable in order: ``setup``,
|
||||||
|
``run``, and ``report``. In this case no options are passed to any of the
|
||||||
|
executables and no outputs from running them are checked. Consequently, the
|
||||||
|
implementation could be simplified with a for-loop as follows:
|
||||||
|
|
||||||
|
.. code-block:: python
|
||||||
|
|
||||||
|
class MyPackage(Package):
|
||||||
|
...
|
||||||
|
|
||||||
|
def test_series(self):
|
||||||
|
"""execute series setup, run, and report"""
|
||||||
|
|
||||||
|
for exe, reason in [
|
||||||
|
("setup", "setup operation"),
|
||||||
|
("run", "perform operation"),
|
||||||
|
("report", "generate report")
|
||||||
|
]:
|
||||||
|
with test_part(self, f"test_series_{exe}", purpose=reason):
|
||||||
|
exe = which(self.prefix.bin.join(exe))
|
||||||
|
exe()
|
||||||
|
|
||||||
|
In both cases, since we're using a context manager, each test part in
|
||||||
|
``test_series`` will execute regardless of the status of the other test
|
||||||
|
parts.
|
||||||
|
|
||||||
|
Now let's look at the output from running the stand-alone tests where
|
||||||
|
the second test part, ``test_series_run``, fails.
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
@@ -5410,50 +5536,68 @@ is illustrated below.
|
|||||||
$ spack test results -l mypackage
|
$ spack test results -l mypackage
|
||||||
==> Results for test suite 'mypackage':
|
==> Results for test suite 'mypackage':
|
||||||
...
|
...
|
||||||
==> [2023-03-10-16:03:56.625204] test: test_example: run installed examples
|
==> [2024-03-10-16:03:56.625204] test: test_series: execute series setup, run, and report
|
||||||
==> [2023-03-10-16:03:56.625439] test: test_example_ex1: run installed ex1
|
==> [2024-03-10-16:03:56.625439] test: test_series_setup: setup operation
|
||||||
...
|
...
|
||||||
FAILED
|
PASSED: MyPackage::test_series_setup
|
||||||
==> [2023-03-10-16:03:56.625555] test: test_example_ex2: run installed ex2
|
==> [2024-03-10-16:03:56.625555] test: test_series_run: perform operation
|
||||||
...
|
...
|
||||||
PASSED
|
FAILED: MyPackage::test_series_run
|
||||||
|
==> [2024-03-10-16:03:57.003456] test: test_series_report: generate report
|
||||||
|
...
|
||||||
|
FAILED: MyPackage::test_series_report
|
||||||
|
FAILED: MyPackage::test_series
|
||||||
...
|
...
|
||||||
|
|
||||||
.. warning::
|
Since test parts depended on the success of previous parts, we see that the
|
||||||
|
failure of one results in the failure of subsequent checks and the overall
|
||||||
|
result of the test method, ``test_series``, is failure.
|
||||||
|
|
||||||
Test results reporting requires that each test method and embedded
|
.. tip::
|
||||||
test part for a package have a unique name.
|
|
||||||
|
|
||||||
Stand-alone tests run in an environment that provides access to information
|
If you want to see more examples from packages using ``test_part``, run
|
||||||
Spack has on how the software was built, such as build options, dependencies,
|
``spack pkg grep "test_part(" | sed "s/\/package.py.*//g" | sort -u``
|
||||||
and compilers. Build options and dependencies are accessed with the normal
|
from the command line to get a list of the packages.
|
||||||
spec checks. Examples of checking :ref:`variant settings <variants>` and
|
|
||||||
:ref:`spec constraints <testing-specs>` can be found at the provided links.
|
|
||||||
Accessing compilers in stand-alone tests that are used by the build requires
|
|
||||||
setting a package property as described :ref:`below <test-compilation>`.
|
|
||||||
|
|
||||||
|
.. _test-build-tests:
|
||||||
|
|
||||||
.. _test-compilation:
|
"""""""""""""""""""""""""""""""""""""
|
||||||
|
Building and running test executables
|
||||||
|
"""""""""""""""""""""""""""""""""""""
|
||||||
|
|
||||||
"""""""""""""""""""""""""
|
.. admonition:: Re-use build-time sources and (small) input data sets when possible.
|
||||||
Enabling test compilation
|
|
||||||
"""""""""""""""""""""""""
|
|
||||||
|
|
||||||
If you want to build and run binaries in tests, then you'll need to tell
|
We **highly recommend** re-using build-time test sources and pared down
|
||||||
Spack to load the package's compiler configuration. This is accomplished
|
input files for testing installed software. These files are easier
|
||||||
by setting the package's ``test_requires_compiler`` property to ``True``.
|
to keep synchronized with software capabilities when they reside
|
||||||
|
within the software's repository. More information on saving files from
|
||||||
|
the installation process can be found at :ref:`cache_extra_test_sources`.
|
||||||
|
|
||||||
Setting the property to ``True`` ensures access to the compiler through
|
If that is not possible, you can add test-related files to the package
|
||||||
canonical environment variables (e.g., ``CC``, ``CXX``, ``FC``, ``F77``).
|
repository (see :ref:`cache_custom_files`). It will be important to
|
||||||
It also gives access to build dependencies like ``cmake`` through their
|
remember to maintain them so they work across listed or supported versions
|
||||||
``spec objects`` (e.g., ``self.spec["cmake"].prefix.bin.cmake``).
|
of the package.
|
||||||
|
|
||||||
.. note::
|
Packages that build libraries are good examples of cases where you'll want
|
||||||
|
to build test executables from the installed software before running them.
|
||||||
|
Doing so requires you to let Spack know it needs to load the package's
|
||||||
|
compiler configuration. This is accomplished by setting the package's
|
||||||
|
``test_requires_compiler`` property to ``True``.
|
||||||
|
|
||||||
The ``test_requires_compiler`` property should be added at the top of
|
.. admonition:: ``test_requires_compiler = True`` is required to build test executables.
|
||||||
the package near other attributes, such as the ``homepage`` and ``url``.
|
|
||||||
|
|
||||||
Below illustrates using this feature to compile an example.
|
Setting the property to ``True`` ensures access to the compiler through
|
||||||
|
canonical environment variables (e.g., ``CC``, ``CXX``, ``FC``, ``F77``).
|
||||||
|
It also gives access to build dependencies like ``cmake`` through their
|
||||||
|
``spec objects`` (e.g., ``self.spec["cmake"].prefix.bin.cmake`` for the
|
||||||
|
path or ``self.spec["cmake"].command`` for the ``Executable`` instance).
|
||||||
|
|
||||||
|
Be sure to add the property at the top of the package class under other
|
||||||
|
properties like the ``homepage``.
|
||||||
|
|
||||||
|
The example below, which ignores how ``cxx-example.cpp`` is acquired,
|
||||||
|
illustrates the basic process of compiling a test executable using the
|
||||||
|
installed library before running it.
|
||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
@@ -5477,28 +5621,22 @@ Below illustrates using this feature to compile an example.
|
|||||||
cxx_example = which(exe)
|
cxx_example = which(exe)
|
||||||
cxx_example()
|
cxx_example()
|
||||||
|
|
||||||
|
Typically the files used to build and or run test executables are either
|
||||||
|
cached from the installation (see :ref:`cache_extra_test_sources`) or added
|
||||||
|
to the package repository (see :ref:`cache_custom_files`). There is nothing
|
||||||
|
preventing the use of both.
|
||||||
|
|
||||||
.. _cache_extra_test_sources:
|
.. _cache_extra_test_sources:
|
||||||
|
|
||||||
"""""""""""""""""""""""
|
""""""""""""""""""""""""""""""""""""
|
||||||
Saving build-time files
|
Saving build- and install-time files
|
||||||
"""""""""""""""""""""""
|
""""""""""""""""""""""""""""""""""""
|
||||||
|
|
||||||
.. note::
|
You can use the ``cache_extra_test_sources`` helper routine to copy
|
||||||
|
directories and or files from the source build stage directory to the
|
||||||
We highly recommend re-using build-time test sources and pared down
|
package's installation directory. Spack will automatically copy these
|
||||||
input files for testing installed software. These files are easier
|
files for you when it sets up the test stage directory and before it
|
||||||
to keep synchronized with software capabilities since they reside
|
begins running the tests.
|
||||||
within the software's repository.
|
|
||||||
|
|
||||||
If that is not possible, you can add test-related files to the package
|
|
||||||
repository (see :ref:`adding custom files <cache_custom_files>`). It
|
|
||||||
will be important to maintain them so they work across listed or supported
|
|
||||||
versions of the package.
|
|
||||||
|
|
||||||
You can use the ``cache_extra_test_sources`` helper to copy directories
|
|
||||||
and or files from the source build stage directory to the package's
|
|
||||||
installation directory.
|
|
||||||
|
|
||||||
The signature for ``cache_extra_test_sources`` is:
|
The signature for ``cache_extra_test_sources`` is:
|
||||||
|
|
||||||
@@ -5513,46 +5651,69 @@ where each argument has the following meaning:
|
|||||||
* ``srcs`` is a string *or* a list of strings corresponding to the
|
* ``srcs`` is a string *or* a list of strings corresponding to the
|
||||||
paths of subdirectories and or files needed for stand-alone testing.
|
paths of subdirectories and or files needed for stand-alone testing.
|
||||||
|
|
||||||
The paths must be relative to the staged source directory. Contents of
|
.. warning::
|
||||||
subdirectories and files are copied to a special test cache subdirectory
|
|
||||||
of the installation prefix. They are automatically copied to the appropriate
|
|
||||||
relative paths under the test stage directory prior to executing stand-alone
|
|
||||||
tests.
|
|
||||||
|
|
||||||
For example, a package method for copying everything in the ``tests``
|
Paths provided in the ``srcs`` argument **must be relative** to the
|
||||||
subdirectory plus the ``foo.c`` and ``bar.c`` files from ``examples``
|
staged source directory. They will be copied to the equivalent relative
|
||||||
and using ``foo.c`` in a test method is illustrated below.
|
location under the test stage directory prior to test execution.
|
||||||
|
|
||||||
|
Contents of subdirectories and files are copied to a special test cache
|
||||||
|
subdirectory of the installation prefix. They are automatically copied to
|
||||||
|
the appropriate relative paths under the test stage directory prior to
|
||||||
|
executing stand-alone tests.
|
||||||
|
|
||||||
|
.. tip::
|
||||||
|
|
||||||
|
*Perform test-related conversions once when copying files.*
|
||||||
|
|
||||||
|
If one or more of the copied files needs to be modified to reference
|
||||||
|
the installed software, it is recommended that those changes be made
|
||||||
|
to the cached files **once** in the post-``install`` copy method
|
||||||
|
**after** the call to ``cache_extra_test_sources``. This will reduce
|
||||||
|
the amount of unnecessary work in the test method **and** avoid problems
|
||||||
|
running stand-alone tests in shared instances and facility deployments.
|
||||||
|
|
||||||
|
The ``filter_file`` function can be quite useful for such changes
|
||||||
|
(see :ref:`file-filtering`).
|
||||||
|
|
||||||
|
Below is a basic example of a test that relies on files from the installation.
|
||||||
|
This package method re-uses the contents of the ``examples`` subdirectory,
|
||||||
|
which is assumed to have all of the files implemented to allow ``make`` to
|
||||||
|
compile and link ``foo.c`` and ``bar.c`` against the package's installed
|
||||||
|
library.
|
||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
class MyLibPackage(Package):
|
class MyLibPackage(MakefilePackage):
|
||||||
...
|
...
|
||||||
|
|
||||||
@run_after("install")
|
@run_after("install")
|
||||||
def copy_test_files(self):
|
def copy_test_files(self):
|
||||||
srcs = ["tests",
|
cache_extra_test_sources(self, "examples")
|
||||||
join_path("examples", "foo.c"),
|
|
||||||
join_path("examples", "bar.c")]
|
|
||||||
cache_extra_test_sources(self, srcs)
|
|
||||||
|
|
||||||
def test_foo(self):
|
def test_example(self):
|
||||||
exe = "foo"
|
"""build and run the examples"""
|
||||||
src_dir = self.test_suite.current_test_cache_dir.examples
|
examples_dir = self.test_suite.current_test_cache_dir.examples
|
||||||
with working_dir(src_dir):
|
with working_dir(examples_dir):
|
||||||
cc = which(os.environ["CC"])
|
make = which("make")
|
||||||
cc(
|
make()
|
||||||
f"-L{self.prefix.lib}",
|
|
||||||
f"-I{self.prefix.include}",
|
|
||||||
f"{exe}.c",
|
|
||||||
"-o", exe
|
|
||||||
)
|
|
||||||
foo = which(exe)
|
|
||||||
foo()
|
|
||||||
|
|
||||||
In this case, the method copies the associated files from the build
|
for program in ["foo", "bar"]:
|
||||||
stage, **after** the software is installed, to the package's test
|
with test_part(
|
||||||
cache directory. Then ``test_foo`` builds ``foo`` using ``foo.c``
|
self,
|
||||||
before running the program.
|
f"test_example_{program}",
|
||||||
|
purpose=f"ensure {program} runs"
|
||||||
|
):
|
||||||
|
exe = Executable(program)
|
||||||
|
exe()
|
||||||
|
|
||||||
|
In this case, ``copy_test_files`` copies the associated files from the
|
||||||
|
build stage to the package's test cache directory under the installation
|
||||||
|
prefix. Running ``spack test run`` for the package results in Spack copying
|
||||||
|
the directory and its contents to the the test stage directory. The
|
||||||
|
``working_dir`` context manager ensures the commands within it are executed
|
||||||
|
from the ``examples_dir``. The test builds the software using ``make`` before
|
||||||
|
running each executable, ``foo`` and ``bar``, as independent test parts.
|
||||||
|
|
||||||
.. note::
|
.. note::
|
||||||
|
|
||||||
@@ -5561,43 +5722,18 @@ before running the program.
|
|||||||
|
|
||||||
The key to copying files for stand-alone testing at build time is use
|
The key to copying files for stand-alone testing at build time is use
|
||||||
of the ``run_after`` directive, which ensures the associated files are
|
of the ``run_after`` directive, which ensures the associated files are
|
||||||
copied **after** the provided build stage where the files **and**
|
copied **after** the provided build stage (``install``) when the installation
|
||||||
installation prefix are available.
|
prefix **and** files are available.
|
||||||
|
|
||||||
These paths are **automatically copied** from cache to the test stage
|
The test method uses the path contained in the package's
|
||||||
directory prior to the execution of any stand-alone tests. Tests access
|
``self.test_suite.current_test_cache_dir`` property for the root directory
|
||||||
the files using the ``self.test_suite.current_test_cache_dir`` property.
|
of the copied files. In this case, that's the ``examples`` subdirectory.
|
||||||
In our example above, test methods can use the following paths to reference
|
|
||||||
the copy of each entry listed in ``srcs``, respectively:
|
|
||||||
|
|
||||||
* ``self.test_suite.current_test_cache_dir.tests``
|
.. tip::
|
||||||
* ``join_path(self.test_suite.current_test_cache_dir.examples, "foo.c")``
|
|
||||||
* ``join_path(self.test_suite.current_test_cache_dir.examples, "bar.c")``
|
|
||||||
|
|
||||||
.. admonition:: Library packages should build stand-alone tests
|
|
||||||
|
|
||||||
Library developers will want to build the associated tests
|
|
||||||
against their **installed** libraries before running them.
|
|
||||||
|
|
||||||
.. note::
|
|
||||||
|
|
||||||
While source and input files are generally recommended, binaries
|
|
||||||
**may** also be cached by the build process. Only you, as the package
|
|
||||||
writer or maintainer, know whether these files would be appropriate
|
|
||||||
for testing the installed software weeks to months later.
|
|
||||||
|
|
||||||
.. note::
|
|
||||||
|
|
||||||
If one or more of the copied files needs to be modified to reference
|
|
||||||
the installed software, it is recommended that those changes be made
|
|
||||||
to the cached files **once** in the ``copy_test_sources`` method and
|
|
||||||
***after** the call to ``cache_extra_test_sources()``. This will
|
|
||||||
reduce the amount of unnecessary work in the test method **and** avoid
|
|
||||||
problems testing in shared instances and facility deployments.
|
|
||||||
|
|
||||||
The ``filter_file`` function can be quite useful for such changes.
|
|
||||||
See :ref:`file manipulation <file-manipulation>`.
|
|
||||||
|
|
||||||
|
If you want to see more examples from packages that cache build files, run
|
||||||
|
``spack pkg grep cache_extra_test_sources | sed "s/\/package.py.*//g" | sort -u``
|
||||||
|
from the command line to get a list of the packages.
|
||||||
|
|
||||||
.. _cache_custom_files:
|
.. _cache_custom_files:
|
||||||
|
|
||||||
@@ -5605,8 +5741,9 @@ the copy of each entry listed in ``srcs``, respectively:
|
|||||||
Adding custom files
|
Adding custom files
|
||||||
"""""""""""""""""""
|
"""""""""""""""""""
|
||||||
|
|
||||||
In some cases it can be useful to have files that can be used to build or
|
Sometimes it is helpful or necessary to include custom files for building and
|
||||||
check the results of tests. Examples include:
|
or checking the results of tests as part of the package. Examples of the types
|
||||||
|
of files that might be useful are:
|
||||||
|
|
||||||
- test source files
|
- test source files
|
||||||
- test input files
|
- test input files
|
||||||
@@ -5614,17 +5751,15 @@ check the results of tests. Examples include:
|
|||||||
- expected test outputs
|
- expected test outputs
|
||||||
|
|
||||||
While obtaining such files from the software repository is preferred (see
|
While obtaining such files from the software repository is preferred (see
|
||||||
:ref:`adding build-time files <cache_extra_test_sources>`), there are
|
:ref:`cache_extra_test_sources`), there are circumstances where doing so is not
|
||||||
circumstances where that is not feasible (e.g., the software is not being
|
feasible such as when the software is not being actively maintained. When test
|
||||||
actively maintained). When test files can't be obtained from the repository
|
files cannot be obtained from the repository or there is a need to supplement
|
||||||
or as a supplement to files that can, Spack supports the inclusion of
|
files that can, Spack supports the inclusion of additional files under the
|
||||||
additional files under the ``test`` subdirectory of the package in the
|
``test`` subdirectory of the package in the Spack repository.
|
||||||
Spack repository.
|
|
||||||
|
|
||||||
Spack **automatically copies** the contents of that directory to the
|
The following example assumes a ``custom-example.c`` is saved in ``MyLibary``
|
||||||
test staging directory prior to running stand-alone tests. Test methods
|
package's ``test`` subdirectory. It also assumes the program simply needs to
|
||||||
access those files using the ``self.test_suite.current_test_data_dir``
|
be compiled and linked against the installed ``MyLibrary`` software.
|
||||||
property as shown below.
|
|
||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
@@ -5634,17 +5769,29 @@ property as shown below.
|
|||||||
test_requires_compiler = True
|
test_requires_compiler = True
|
||||||
...
|
...
|
||||||
|
|
||||||
def test_example(self):
|
def test_custom_example(self):
|
||||||
"""build and run custom-example"""
|
"""build and run custom-example"""
|
||||||
data_dir = self.test_suite.current_test_data_dir
|
src_dir = self.test_suite.current_test_data_dir
|
||||||
exe = "custom-example"
|
exe = "custom-example"
|
||||||
src = datadir.join(f"{exe}.cpp")
|
|
||||||
...
|
|
||||||
# TODO: Build custom-example using src and exe
|
|
||||||
...
|
|
||||||
custom_example = which(exe)
|
|
||||||
custom_example()
|
|
||||||
|
|
||||||
|
with working_dir(src_dir):
|
||||||
|
cc = which(os.environ["CC"])
|
||||||
|
cc(
|
||||||
|
f"-L{self.prefix.lib}",
|
||||||
|
f"-I{self.prefix.include}",
|
||||||
|
f"{exe}.cpp",
|
||||||
|
"-o", exe
|
||||||
|
)
|
||||||
|
|
||||||
|
custom_example = Executable(exe)
|
||||||
|
custom_example()
|
||||||
|
|
||||||
|
In this case, ``spack test run`` for the package results in Spack copying
|
||||||
|
the contents of the ``test`` subdirectory to the test stage directory path
|
||||||
|
in ``self.test_suite.current_test_data_dir`` before calling
|
||||||
|
``test_custom_example``. Use of the ``working_dir`` context manager
|
||||||
|
ensures the commands to build and run the program are performed from
|
||||||
|
within the appropriate subdirectory of the test stage.
|
||||||
|
|
||||||
.. _expected_test_output_from_file:
|
.. _expected_test_output_from_file:
|
||||||
|
|
||||||
@@ -5653,9 +5800,8 @@ Reading expected output from a file
|
|||||||
"""""""""""""""""""""""""""""""""""
|
"""""""""""""""""""""""""""""""""""
|
||||||
|
|
||||||
The helper function ``get_escaped_text_output`` is available for packages
|
The helper function ``get_escaped_text_output`` is available for packages
|
||||||
to retrieve and properly format the text from a file that contains the
|
to retrieve properly formatted text from a file potentially containing
|
||||||
expected output from running an executable that may contain special
|
special characters.
|
||||||
characters.
|
|
||||||
|
|
||||||
The signature for ``get_escaped_text_output`` is:
|
The signature for ``get_escaped_text_output`` is:
|
||||||
|
|
||||||
@@ -5665,10 +5811,13 @@ The signature for ``get_escaped_text_output`` is:
|
|||||||
|
|
||||||
where ``filename`` is the path to the file containing the expected output.
|
where ``filename`` is the path to the file containing the expected output.
|
||||||
|
|
||||||
The ``filename`` for a :ref:`custom file <cache_custom_files>` can be
|
The path provided to ``filename`` for one of the copied custom files
|
||||||
accessed by tests using the ``self.test_suite.current_test_data_dir``
|
(:ref:`custom file <cache_custom_files>`) is in the path rooted at
|
||||||
property. The example below illustrates how to read a file that was
|
``self.test_suite.current_test_data_dir``.
|
||||||
added to the package's ``test`` subdirectory.
|
|
||||||
|
The example below shows how to reference both the custom database
|
||||||
|
(``packages.db``) and expected output (``dump.out``) files Spack copies
|
||||||
|
to the test stage:
|
||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
@@ -5690,8 +5839,9 @@ added to the package's ``test`` subdirectory.
|
|||||||
for exp in expected:
|
for exp in expected:
|
||||||
assert re.search(exp, out), f"Expected '{exp}' in output"
|
assert re.search(exp, out), f"Expected '{exp}' in output"
|
||||||
|
|
||||||
If the file was instead copied from the ``tests`` subdirectory of the staged
|
If the files were instead cached from installing the software, the paths to the
|
||||||
source code, the path would be obtained as shown below.
|
two files would be found under the ``self.test_suite.current_test_cache_dir``
|
||||||
|
directory as shown below:
|
||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
@@ -5699,17 +5849,24 @@ source code, the path would be obtained as shown below.
|
|||||||
"""check example table dump"""
|
"""check example table dump"""
|
||||||
test_cache_dir = self.test_suite.current_test_cache_dir
|
test_cache_dir = self.test_suite.current_test_cache_dir
|
||||||
db_filename = test_cache_dir.join("packages.db")
|
db_filename = test_cache_dir.join("packages.db")
|
||||||
|
..
|
||||||
|
expected = get_escaped_text_output(test_cache_dir.join("dump.out"))
|
||||||
|
...
|
||||||
|
|
||||||
Alternatively, if the file was copied to the ``share/tests`` subdirectory
|
Alternatively, if both files had been installed by the software into the
|
||||||
as part of the installation process, the test could access the path as
|
``share/tests`` subdirectory of the installation prefix, the paths to the
|
||||||
follows:
|
two files would be referenced as follows:
|
||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
def test_example(self):
|
def test_example(self):
|
||||||
"""check example table dump"""
|
"""check example table dump"""
|
||||||
db_filename = join_path(self.prefix.share.tests, "packages.db")
|
db_filename = self.prefix.share.tests.join("packages.db")
|
||||||
|
..
|
||||||
|
expected = get_escaped_text_output(
|
||||||
|
self.prefix.share.tests.join("dump.out")
|
||||||
|
)
|
||||||
|
...
|
||||||
|
|
||||||
.. _check_outputs:
|
.. _check_outputs:
|
||||||
|
|
||||||
@@ -5717,9 +5874,9 @@ follows:
|
|||||||
Comparing expected to actual outputs
|
Comparing expected to actual outputs
|
||||||
""""""""""""""""""""""""""""""""""""
|
""""""""""""""""""""""""""""""""""""
|
||||||
|
|
||||||
The helper function ``check_outputs`` is available for packages to ensure
|
The ``check_outputs`` helper routine is available for packages to ensure
|
||||||
the expected outputs from running an executable are contained within the
|
multiple expected outputs from running an executable are contained within
|
||||||
actual outputs.
|
the actual outputs.
|
||||||
|
|
||||||
The signature for ``check_outputs`` is:
|
The signature for ``check_outputs`` is:
|
||||||
|
|
||||||
@@ -5745,11 +5902,17 @@ Invoking the method is the equivalent of:
|
|||||||
if errors:
|
if errors:
|
||||||
raise RuntimeError("\n ".join(errors))
|
raise RuntimeError("\n ".join(errors))
|
||||||
|
|
||||||
|
.. tip::
|
||||||
|
|
||||||
|
If you want to see more examples from packages that use this helper, run
|
||||||
|
``spack pkg grep check_outputs | sed "s/\/package.py.*//g" | sort -u``
|
||||||
|
from the command line to get a list of the packages.
|
||||||
|
|
||||||
|
|
||||||
.. _accessing-files:
|
.. _accessing-files:
|
||||||
|
|
||||||
"""""""""""""""""""""""""""""""""""""""""
|
"""""""""""""""""""""""""""""""""""""""""
|
||||||
Accessing package- and test-related files
|
Finding package- and test-related files
|
||||||
"""""""""""""""""""""""""""""""""""""""""
|
"""""""""""""""""""""""""""""""""""""""""
|
||||||
|
|
||||||
You may need to access files from one or more locations when writing
|
You may need to access files from one or more locations when writing
|
||||||
@@ -5758,8 +5921,7 @@ include test source files or includes them but has no way to build the
|
|||||||
executables using the installed headers and libraries. In these cases
|
executables using the installed headers and libraries. In these cases
|
||||||
you may need to reference the files relative to one or more root directory.
|
you may need to reference the files relative to one or more root directory.
|
||||||
The table below lists relevant path properties and provides additional
|
The table below lists relevant path properties and provides additional
|
||||||
examples of their use.
|
examples of their use. See :ref:`expected_test_output_from_file` for
|
||||||
:ref:`Reading expected output <expected_test_output_from_file>` provides
|
|
||||||
examples of accessing files saved from the software repository, package
|
examples of accessing files saved from the software repository, package
|
||||||
repository, and installation.
|
repository, and installation.
|
||||||
|
|
||||||
@@ -5788,7 +5950,6 @@ repository, and installation.
|
|||||||
- ``self.test_suite.current_test_data_dir``
|
- ``self.test_suite.current_test_data_dir``
|
||||||
- ``join_path(self.test_suite.current_test_data_dir, "hello.f90")``
|
- ``join_path(self.test_suite.current_test_data_dir, "hello.f90")``
|
||||||
|
|
||||||
|
|
||||||
.. _inheriting-tests:
|
.. _inheriting-tests:
|
||||||
|
|
||||||
""""""""""""""""""""""""""""
|
""""""""""""""""""""""""""""
|
||||||
@@ -5831,7 +5992,7 @@ maintainers provide additional stand-alone tests customized to the package.
|
|||||||
.. warning::
|
.. warning::
|
||||||
|
|
||||||
Any package that implements a test method with the same name as an
|
Any package that implements a test method with the same name as an
|
||||||
inherited method overrides the inherited method. If that is not the
|
inherited method will override the inherited method. If that is not the
|
||||||
goal and you are not explicitly calling and adding functionality to
|
goal and you are not explicitly calling and adding functionality to
|
||||||
the inherited method for the test, then make sure that all test methods
|
the inherited method for the test, then make sure that all test methods
|
||||||
and embedded test parts have unique test names.
|
and embedded test parts have unique test names.
|
||||||
@@ -5996,6 +6157,8 @@ running:
|
|||||||
This is already part of the boilerplate for packages created with
|
This is already part of the boilerplate for packages created with
|
||||||
``spack create``.
|
``spack create``.
|
||||||
|
|
||||||
|
.. _file-filtering:
|
||||||
|
|
||||||
^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^
|
||||||
Filtering functions
|
Filtering functions
|
||||||
^^^^^^^^^^^^^^^^^^^
|
^^^^^^^^^^^^^^^^^^^
|
||||||
|
|||||||
@@ -253,17 +253,6 @@ can easily happen if it is not updated frequently, this behavior ensures that
|
|||||||
spack has a way to know for certain about the status of any concrete spec on
|
spack has a way to know for certain about the status of any concrete spec on
|
||||||
the remote mirror, but can slow down pipeline generation significantly.
|
the remote mirror, but can slow down pipeline generation significantly.
|
||||||
|
|
||||||
The ``--optimize`` argument is experimental and runs the generated pipeline
|
|
||||||
document through a series of optimization passes designed to reduce the size
|
|
||||||
of the generated file.
|
|
||||||
|
|
||||||
The ``--dependencies`` is also experimental and disables what in Gitlab is
|
|
||||||
referred to as DAG scheduling, internally using the ``dependencies`` keyword
|
|
||||||
rather than ``needs`` to list dependency jobs. The drawback of using this option
|
|
||||||
is that before any job can begin, all jobs in previous stages must first
|
|
||||||
complete. The benefit is that Gitlab allows more dependencies to be listed
|
|
||||||
when using ``dependencies`` instead of ``needs``.
|
|
||||||
|
|
||||||
The optional ``--output-file`` argument should be an absolute path (including
|
The optional ``--output-file`` argument should be an absolute path (including
|
||||||
file name) to the generated pipeline, and if not given, the default is
|
file name) to the generated pipeline, and if not given, the default is
|
||||||
``./.gitlab-ci.yml``.
|
``./.gitlab-ci.yml``.
|
||||||
|
|||||||
@@ -476,9 +476,3 @@ implemented using Python's built-in `sys.path
|
|||||||
:py:mod:`spack.repo` module implements a custom `Python importer
|
:py:mod:`spack.repo` module implements a custom `Python importer
|
||||||
<https://docs.python.org/2/library/imp.html>`_.
|
<https://docs.python.org/2/library/imp.html>`_.
|
||||||
|
|
||||||
.. warning::
|
|
||||||
|
|
||||||
The mechanism for extending packages is not yet extensively tested,
|
|
||||||
and extending packages across repositories imposes inter-repo
|
|
||||||
dependencies, which may be hard to manage. Use this feature at your
|
|
||||||
own risk, but let us know if you have a use case for it.
|
|
||||||
|
|||||||
@@ -1,13 +1,13 @@
|
|||||||
sphinx==7.2.6
|
sphinx==7.4.7
|
||||||
sphinxcontrib-programoutput==0.17
|
sphinxcontrib-programoutput==0.17
|
||||||
sphinx_design==0.5.0
|
sphinx_design==0.6.0
|
||||||
sphinx-rtd-theme==2.0.0
|
sphinx-rtd-theme==2.0.0
|
||||||
python-levenshtein==0.25.1
|
python-levenshtein==0.25.1
|
||||||
docutils==0.20.1
|
docutils==0.20.1
|
||||||
pygments==2.17.2
|
pygments==2.18.0
|
||||||
urllib3==2.2.1
|
urllib3==2.2.2
|
||||||
pytest==8.2.0
|
pytest==8.3.1
|
||||||
isort==5.13.2
|
isort==5.13.2
|
||||||
black==24.4.2
|
black==24.4.2
|
||||||
flake8==7.0.0
|
flake8==7.1.0
|
||||||
mypy==1.10.0
|
mypy==1.11.0
|
||||||
|
|||||||
@@ -98,3 +98,10 @@ def path_filter_caller(*args, **kwargs):
|
|||||||
if _func:
|
if _func:
|
||||||
return holder_func(_func)
|
return holder_func(_func)
|
||||||
return holder_func
|
return holder_func
|
||||||
|
|
||||||
|
|
||||||
|
def sanitize_win_longpath(path: str) -> str:
|
||||||
|
"""Strip Windows extended path prefix from strings
|
||||||
|
Returns sanitized string.
|
||||||
|
no-op if extended path prefix is not present"""
|
||||||
|
return path.lstrip("\\\\?\\")
|
||||||
|
|||||||
@@ -187,12 +187,18 @@ def polite_filename(filename: str) -> str:
|
|||||||
return _polite_antipattern().sub("_", filename)
|
return _polite_antipattern().sub("_", filename)
|
||||||
|
|
||||||
|
|
||||||
def getuid():
|
def getuid() -> Union[str, int]:
|
||||||
|
"""Returns os getuid on non Windows
|
||||||
|
On Windows returns 0 for admin users, login string otherwise
|
||||||
|
This is in line with behavior from get_owner_uid which
|
||||||
|
always returns the login string on Windows
|
||||||
|
"""
|
||||||
if sys.platform == "win32":
|
if sys.platform == "win32":
|
||||||
import ctypes
|
import ctypes
|
||||||
|
|
||||||
|
# If not admin, use the string name of the login as a unique ID
|
||||||
if ctypes.windll.shell32.IsUserAnAdmin() == 0:
|
if ctypes.windll.shell32.IsUserAnAdmin() == 0:
|
||||||
return 1
|
return os.getlogin()
|
||||||
return 0
|
return 0
|
||||||
else:
|
else:
|
||||||
return os.getuid()
|
return os.getuid()
|
||||||
@@ -213,6 +219,15 @@ def _win_rename(src, dst):
|
|||||||
os.replace(src, dst)
|
os.replace(src, dst)
|
||||||
|
|
||||||
|
|
||||||
|
@system_path_filter
|
||||||
|
def msdos_escape_parens(path):
|
||||||
|
"""MS-DOS interprets parens as grouping parameters even in a quoted string"""
|
||||||
|
if sys.platform == "win32":
|
||||||
|
return path.replace("(", "^(").replace(")", "^)")
|
||||||
|
else:
|
||||||
|
return path
|
||||||
|
|
||||||
|
|
||||||
@system_path_filter
|
@system_path_filter
|
||||||
def rename(src, dst):
|
def rename(src, dst):
|
||||||
# On Windows, os.rename will fail if the destination file already exists
|
# On Windows, os.rename will fail if the destination file already exists
|
||||||
@@ -553,7 +568,13 @@ def exploding_archive_handler(tarball_container, stage):
|
|||||||
|
|
||||||
|
|
||||||
@system_path_filter(arg_slice=slice(1))
|
@system_path_filter(arg_slice=slice(1))
|
||||||
def get_owner_uid(path, err_msg=None):
|
def get_owner_uid(path, err_msg=None) -> Union[str, int]:
|
||||||
|
"""Returns owner UID of path destination
|
||||||
|
On non Windows this is the value of st_uid
|
||||||
|
On Windows this is the login string associated with the
|
||||||
|
owning user.
|
||||||
|
|
||||||
|
"""
|
||||||
if not os.path.exists(path):
|
if not os.path.exists(path):
|
||||||
mkdirp(path, mode=stat.S_IRWXU)
|
mkdirp(path, mode=stat.S_IRWXU)
|
||||||
|
|
||||||
@@ -745,7 +766,6 @@ def copy_tree(
|
|||||||
src: str,
|
src: str,
|
||||||
dest: str,
|
dest: str,
|
||||||
symlinks: bool = True,
|
symlinks: bool = True,
|
||||||
allow_broken_symlinks: bool = sys.platform != "win32",
|
|
||||||
ignore: Optional[Callable[[str], bool]] = None,
|
ignore: Optional[Callable[[str], bool]] = None,
|
||||||
_permissions: bool = False,
|
_permissions: bool = False,
|
||||||
):
|
):
|
||||||
@@ -768,8 +788,6 @@ def copy_tree(
|
|||||||
src (str): the directory to copy
|
src (str): the directory to copy
|
||||||
dest (str): the destination directory
|
dest (str): the destination directory
|
||||||
symlinks (bool): whether or not to preserve symlinks
|
symlinks (bool): whether or not to preserve symlinks
|
||||||
allow_broken_symlinks (bool): whether or not to allow broken (dangling) symlinks,
|
|
||||||
On Windows, setting this to True will raise an exception. Defaults to true on unix.
|
|
||||||
ignore (typing.Callable): function indicating which files to ignore
|
ignore (typing.Callable): function indicating which files to ignore
|
||||||
_permissions (bool): for internal use only
|
_permissions (bool): for internal use only
|
||||||
|
|
||||||
@@ -777,8 +795,6 @@ def copy_tree(
|
|||||||
IOError: if *src* does not match any files or directories
|
IOError: if *src* does not match any files or directories
|
||||||
ValueError: if *src* is a parent directory of *dest*
|
ValueError: if *src* is a parent directory of *dest*
|
||||||
"""
|
"""
|
||||||
if allow_broken_symlinks and sys.platform == "win32":
|
|
||||||
raise llnl.util.symlink.SymlinkError("Cannot allow broken symlinks on Windows!")
|
|
||||||
if _permissions:
|
if _permissions:
|
||||||
tty.debug("Installing {0} to {1}".format(src, dest))
|
tty.debug("Installing {0} to {1}".format(src, dest))
|
||||||
else:
|
else:
|
||||||
@@ -822,7 +838,7 @@ def copy_tree(
|
|||||||
if islink(s):
|
if islink(s):
|
||||||
link_target = resolve_link_target_relative_to_the_link(s)
|
link_target = resolve_link_target_relative_to_the_link(s)
|
||||||
if symlinks:
|
if symlinks:
|
||||||
target = os.readlink(s)
|
target = readlink(s)
|
||||||
if os.path.isabs(target):
|
if os.path.isabs(target):
|
||||||
|
|
||||||
def escaped_path(path):
|
def escaped_path(path):
|
||||||
@@ -851,16 +867,14 @@ def escaped_path(path):
|
|||||||
copy_mode(s, d)
|
copy_mode(s, d)
|
||||||
|
|
||||||
for target, d, s in links:
|
for target, d, s in links:
|
||||||
symlink(target, d, allow_broken_symlinks=allow_broken_symlinks)
|
symlink(target, d)
|
||||||
if _permissions:
|
if _permissions:
|
||||||
set_install_permissions(d)
|
set_install_permissions(d)
|
||||||
copy_mode(s, d)
|
copy_mode(s, d)
|
||||||
|
|
||||||
|
|
||||||
@system_path_filter
|
@system_path_filter
|
||||||
def install_tree(
|
def install_tree(src, dest, symlinks=True, ignore=None):
|
||||||
src, dest, symlinks=True, ignore=None, allow_broken_symlinks=sys.platform != "win32"
|
|
||||||
):
|
|
||||||
"""Recursively install an entire directory tree rooted at *src*.
|
"""Recursively install an entire directory tree rooted at *src*.
|
||||||
|
|
||||||
Same as :py:func:`copy_tree` with the addition of setting proper
|
Same as :py:func:`copy_tree` with the addition of setting proper
|
||||||
@@ -871,21 +885,12 @@ def install_tree(
|
|||||||
dest (str): the destination directory
|
dest (str): the destination directory
|
||||||
symlinks (bool): whether or not to preserve symlinks
|
symlinks (bool): whether or not to preserve symlinks
|
||||||
ignore (typing.Callable): function indicating which files to ignore
|
ignore (typing.Callable): function indicating which files to ignore
|
||||||
allow_broken_symlinks (bool): whether or not to allow broken (dangling) symlinks,
|
|
||||||
On Windows, setting this to True will raise an exception.
|
|
||||||
|
|
||||||
Raises:
|
Raises:
|
||||||
IOError: if *src* does not match any files or directories
|
IOError: if *src* does not match any files or directories
|
||||||
ValueError: if *src* is a parent directory of *dest*
|
ValueError: if *src* is a parent directory of *dest*
|
||||||
"""
|
"""
|
||||||
copy_tree(
|
copy_tree(src, dest, symlinks=symlinks, ignore=ignore, _permissions=True)
|
||||||
src,
|
|
||||||
dest,
|
|
||||||
symlinks=symlinks,
|
|
||||||
allow_broken_symlinks=allow_broken_symlinks,
|
|
||||||
ignore=ignore,
|
|
||||||
_permissions=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@system_path_filter
|
@system_path_filter
|
||||||
@@ -2429,9 +2434,10 @@ def add_library_dependent(self, *dest):
|
|||||||
"""
|
"""
|
||||||
for pth in dest:
|
for pth in dest:
|
||||||
if os.path.isfile(pth):
|
if os.path.isfile(pth):
|
||||||
self._additional_library_dependents.add(pathlib.Path(pth).parent)
|
new_pth = pathlib.Path(pth).parent
|
||||||
else:
|
else:
|
||||||
self._additional_library_dependents.add(pathlib.Path(pth))
|
new_pth = pathlib.Path(pth)
|
||||||
|
self._additional_library_dependents.add(new_pth)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def rpaths(self):
|
def rpaths(self):
|
||||||
@@ -2509,8 +2515,14 @@ def establish_link(self):
|
|||||||
|
|
||||||
# for each binary install dir in self.pkg (i.e. pkg.prefix.bin, pkg.prefix.lib)
|
# for each binary install dir in self.pkg (i.e. pkg.prefix.bin, pkg.prefix.lib)
|
||||||
# install a symlink to each dependent library
|
# install a symlink to each dependent library
|
||||||
for library, lib_dir in itertools.product(self.rpaths, self.library_dependents):
|
|
||||||
self._link(library, lib_dir)
|
# do not rpath for system libraries included in the dag
|
||||||
|
# we should not be modifying libraries managed by the Windows system
|
||||||
|
# as this will negatively impact linker behavior and can result in permission
|
||||||
|
# errors if those system libs are not modifiable by Spack
|
||||||
|
if "windows-system" not in getattr(self.pkg, "tags", []):
|
||||||
|
for library, lib_dir in itertools.product(self.rpaths, self.library_dependents):
|
||||||
|
self._link(library, lib_dir)
|
||||||
|
|
||||||
|
|
||||||
@system_path_filter
|
@system_path_filter
|
||||||
|
|||||||
@@ -8,100 +8,75 @@
|
|||||||
import subprocess
|
import subprocess
|
||||||
import sys
|
import sys
|
||||||
import tempfile
|
import tempfile
|
||||||
|
from typing import Union
|
||||||
|
|
||||||
from llnl.util import lang, tty
|
from llnl.util import lang, tty
|
||||||
|
|
||||||
from ..path import system_path_filter
|
from ..path import sanitize_win_longpath, system_path_filter
|
||||||
|
|
||||||
if sys.platform == "win32":
|
if sys.platform == "win32":
|
||||||
from win32file import CreateHardLink
|
from win32file import CreateHardLink
|
||||||
|
|
||||||
is_windows = sys.platform == "win32"
|
|
||||||
|
|
||||||
|
def _windows_symlink(
|
||||||
|
src: str, dst: str, target_is_directory: bool = False, *, dir_fd: Union[int, None] = None
|
||||||
|
):
|
||||||
|
"""On Windows with System Administrator privileges this will be a normal symbolic link via
|
||||||
|
os.symlink. On Windows without privledges the link will be a junction for a directory and a
|
||||||
|
hardlink for a file. On Windows the various link types are:
|
||||||
|
|
||||||
def symlink(source_path: str, link_path: str, allow_broken_symlinks: bool = not is_windows):
|
Symbolic Link: A link to a file or directory on the same or different volume (drive letter) or
|
||||||
"""
|
even to a remote file or directory (using UNC in its path). Need System Administrator
|
||||||
Create a link.
|
privileges to make these.
|
||||||
|
|
||||||
On non-Windows and Windows with System Administrator
|
Hard Link: A link to a file on the same volume (drive letter) only. Every file (file's data)
|
||||||
privleges this will be a normal symbolic link via
|
has at least 1 hard link (file's name). But when this method creates a new hard link there will
|
||||||
os.symlink.
|
be 2. Deleting all hard links effectively deletes the file. Don't need System Administrator
|
||||||
|
privileges.
|
||||||
|
|
||||||
On Windows without privledges the link will be a
|
Junction: A link to a directory on the same or different volume (drive letter) but not to a
|
||||||
junction for a directory and a hardlink for a file.
|
remote directory. Don't need System Administrator privileges."""
|
||||||
On Windows the various link types are:
|
source_path = os.path.normpath(src)
|
||||||
|
|
||||||
Symbolic Link: A link to a file or directory on the
|
|
||||||
same or different volume (drive letter) or even to
|
|
||||||
a remote file or directory (using UNC in its path).
|
|
||||||
Need System Administrator privileges to make these.
|
|
||||||
|
|
||||||
Hard Link: A link to a file on the same volume (drive
|
|
||||||
letter) only. Every file (file's data) has at least 1
|
|
||||||
hard link (file's name). But when this method creates
|
|
||||||
a new hard link there will be 2. Deleting all hard
|
|
||||||
links effectively deletes the file. Don't need System
|
|
||||||
Administrator privileges.
|
|
||||||
|
|
||||||
Junction: A link to a directory on the same or different
|
|
||||||
volume (drive letter) but not to a remote directory. Don't
|
|
||||||
need System Administrator privileges.
|
|
||||||
|
|
||||||
Parameters:
|
|
||||||
source_path (str): The real file or directory that the link points to.
|
|
||||||
Must be absolute OR relative to the link.
|
|
||||||
link_path (str): The path where the link will exist.
|
|
||||||
allow_broken_symlinks (bool): On Linux or Mac, don't raise an exception if the source_path
|
|
||||||
doesn't exist. This will still raise an exception on Windows.
|
|
||||||
"""
|
|
||||||
source_path = os.path.normpath(source_path)
|
|
||||||
win_source_path = source_path
|
win_source_path = source_path
|
||||||
link_path = os.path.normpath(link_path)
|
link_path = os.path.normpath(dst)
|
||||||
|
|
||||||
# Never allow broken links on Windows.
|
# Perform basic checks to make sure symlinking will succeed
|
||||||
if sys.platform == "win32" and allow_broken_symlinks:
|
if os.path.lexists(link_path):
|
||||||
raise ValueError("allow_broken_symlinks parameter cannot be True on Windows.")
|
raise AlreadyExistsError(f"Link path ({link_path}) already exists. Cannot create link.")
|
||||||
|
|
||||||
if not allow_broken_symlinks:
|
if not os.path.exists(source_path):
|
||||||
# Perform basic checks to make sure symlinking will succeed
|
if os.path.isabs(source_path):
|
||||||
if os.path.lexists(link_path):
|
# An absolute source path that does not exist will result in a broken link.
|
||||||
raise AlreadyExistsError(
|
raise SymlinkError(
|
||||||
f"Link path ({link_path}) already exists. Cannot create link."
|
f"Source path ({source_path}) is absolute but does not exist. Resulting "
|
||||||
|
f"link would be broken so not making link."
|
||||||
)
|
)
|
||||||
|
else:
|
||||||
if not os.path.exists(source_path):
|
# os.symlink can create a link when the given source path is relative to
|
||||||
if os.path.isabs(source_path) and not allow_broken_symlinks:
|
# the link path. Emulate this behavior and check to see if the source exists
|
||||||
# An absolute source path that does not exist will result in a broken link.
|
# relative to the link path ahead of link creation to prevent broken
|
||||||
raise SymlinkError(
|
# links from being made.
|
||||||
f"Source path ({source_path}) is absolute but does not exist. Resulting "
|
link_parent_dir = os.path.dirname(link_path)
|
||||||
f"link would be broken so not making link."
|
relative_path = os.path.join(link_parent_dir, source_path)
|
||||||
)
|
if os.path.exists(relative_path):
|
||||||
|
# In order to work on windows, the source path needs to be modified to be
|
||||||
|
# relative because hardlink/junction dont resolve relative paths the same
|
||||||
|
# way as os.symlink. This is ignored on other operating systems.
|
||||||
|
win_source_path = relative_path
|
||||||
else:
|
else:
|
||||||
# os.symlink can create a link when the given source path is relative to
|
raise SymlinkError(
|
||||||
# the link path. Emulate this behavior and check to see if the source exists
|
f"The source path ({source_path}) is not relative to the link path "
|
||||||
# relative to the link path ahead of link creation to prevent broken
|
f"({link_path}). Resulting link would be broken so not making link."
|
||||||
# links from being made.
|
)
|
||||||
link_parent_dir = os.path.dirname(link_path)
|
|
||||||
relative_path = os.path.join(link_parent_dir, source_path)
|
|
||||||
if os.path.exists(relative_path):
|
|
||||||
# In order to work on windows, the source path needs to be modified to be
|
|
||||||
# relative because hardlink/junction dont resolve relative paths the same
|
|
||||||
# way as os.symlink. This is ignored on other operating systems.
|
|
||||||
win_source_path = relative_path
|
|
||||||
elif not allow_broken_symlinks:
|
|
||||||
raise SymlinkError(
|
|
||||||
f"The source path ({source_path}) is not relative to the link path "
|
|
||||||
f"({link_path}). Resulting link would be broken so not making link."
|
|
||||||
)
|
|
||||||
|
|
||||||
# Create the symlink
|
# Create the symlink
|
||||||
if sys.platform == "win32" and not _windows_can_symlink():
|
if not _windows_can_symlink():
|
||||||
_windows_create_link(win_source_path, link_path)
|
_windows_create_link(win_source_path, link_path)
|
||||||
else:
|
else:
|
||||||
os.symlink(source_path, link_path, target_is_directory=os.path.isdir(source_path))
|
os.symlink(source_path, link_path, target_is_directory=os.path.isdir(source_path))
|
||||||
|
|
||||||
|
|
||||||
def islink(path: str) -> bool:
|
def _windows_islink(path: str) -> bool:
|
||||||
"""Override os.islink to give correct answer for spack logic.
|
"""Override os.islink to give correct answer for spack logic.
|
||||||
|
|
||||||
For Non-Windows: a link can be determined with the os.path.islink method.
|
For Non-Windows: a link can be determined with the os.path.islink method.
|
||||||
@@ -247,9 +222,9 @@ def _windows_create_junction(source: str, link: str):
|
|||||||
out, err = proc.communicate()
|
out, err = proc.communicate()
|
||||||
tty.debug(out.decode())
|
tty.debug(out.decode())
|
||||||
if proc.returncode != 0:
|
if proc.returncode != 0:
|
||||||
err = err.decode()
|
err_str = err.decode()
|
||||||
tty.error(err)
|
tty.error(err_str)
|
||||||
raise SymlinkError("Make junction command returned a non-zero return code.", err)
|
raise SymlinkError("Make junction command returned a non-zero return code.", err_str)
|
||||||
|
|
||||||
|
|
||||||
def _windows_create_hard_link(path: str, link: str):
|
def _windows_create_hard_link(path: str, link: str):
|
||||||
@@ -269,14 +244,14 @@ def _windows_create_hard_link(path: str, link: str):
|
|||||||
CreateHardLink(link, path)
|
CreateHardLink(link, path)
|
||||||
|
|
||||||
|
|
||||||
def readlink(path: str):
|
def _windows_readlink(path: str, *, dir_fd=None):
|
||||||
"""Spack utility to override of os.readlink method to work cross platform"""
|
"""Spack utility to override of os.readlink method to work cross platform"""
|
||||||
if _windows_is_hardlink(path):
|
if _windows_is_hardlink(path):
|
||||||
return _windows_read_hard_link(path)
|
return _windows_read_hard_link(path)
|
||||||
elif _windows_is_junction(path):
|
elif _windows_is_junction(path):
|
||||||
return _windows_read_junction(path)
|
return _windows_read_junction(path)
|
||||||
else:
|
else:
|
||||||
return os.readlink(path)
|
return sanitize_win_longpath(os.readlink(path, dir_fd=dir_fd))
|
||||||
|
|
||||||
|
|
||||||
def _windows_read_hard_link(link: str) -> str:
|
def _windows_read_hard_link(link: str) -> str:
|
||||||
@@ -338,6 +313,16 @@ def resolve_link_target_relative_to_the_link(link):
|
|||||||
return os.path.join(link_dir, target)
|
return os.path.join(link_dir, target)
|
||||||
|
|
||||||
|
|
||||||
|
if sys.platform == "win32":
|
||||||
|
symlink = _windows_symlink
|
||||||
|
readlink = _windows_readlink
|
||||||
|
islink = _windows_islink
|
||||||
|
else:
|
||||||
|
symlink = os.symlink
|
||||||
|
readlink = os.readlink
|
||||||
|
islink = os.path.islink
|
||||||
|
|
||||||
|
|
||||||
class SymlinkError(RuntimeError):
|
class SymlinkError(RuntimeError):
|
||||||
"""Exception class for errors raised while creating symlinks,
|
"""Exception class for errors raised while creating symlinks,
|
||||||
junctions and hard links
|
junctions and hard links
|
||||||
|
|||||||
@@ -33,8 +33,23 @@
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
esc, bell, lbracket, bslash, newline = r"\x1b", r"\x07", r"\[", r"\\", r"\n"
|
||||||
|
# Ansi Control Sequence Introducers (CSI) are a well-defined format
|
||||||
|
# Standard ECMA-48: Control Functions for Character-Imaging I/O Devices, section 5.4
|
||||||
|
# https://www.ecma-international.org/wp-content/uploads/ECMA-48_5th_edition_june_1991.pdf
|
||||||
|
csi_pre = f"{esc}{lbracket}"
|
||||||
|
csi_param, csi_inter, csi_post = r"[0-?]", r"[ -/]", r"[@-~]"
|
||||||
|
ansi_csi = f"{csi_pre}{csi_param}*{csi_inter}*{csi_post}"
|
||||||
|
# General ansi escape sequences have well-defined prefixes,
|
||||||
|
# but content and suffixes are less reliable.
|
||||||
|
# Conservatively assume they end with either "<ESC>\" or "<BELL>",
|
||||||
|
# with no intervening "<ESC>"/"<BELL>" keys or newlines
|
||||||
|
esc_pre = f"{esc}[@-_]"
|
||||||
|
esc_content = f"[^{esc}{bell}{newline}]"
|
||||||
|
esc_post = f"(?:{esc}{bslash}|{bell})"
|
||||||
|
ansi_esc = f"{esc_pre}{esc_content}*{esc_post}"
|
||||||
# Use this to strip escape sequences
|
# Use this to strip escape sequences
|
||||||
_escape = re.compile(r"\x1b[^m]*m|\x1b\[?1034h|\x1b\][0-9]+;[^\x07]*\x07")
|
_escape = re.compile(f"{ansi_csi}|{ansi_esc}")
|
||||||
|
|
||||||
# control characters for enabling/disabling echo
|
# control characters for enabling/disabling echo
|
||||||
#
|
#
|
||||||
|
|||||||
@@ -4,7 +4,7 @@
|
|||||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
||||||
|
|
||||||
#: PEP440 canonical <major>.<minor>.<micro>.<devN> string
|
#: PEP440 canonical <major>.<minor>.<micro>.<devN> string
|
||||||
__version__ = "0.22.0.dev0"
|
__version__ = "0.23.0.dev0"
|
||||||
spack_version = __version__
|
spack_version = __version__
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -254,8 +254,8 @@ def _search_duplicate_specs_in_externals(error_cls):
|
|||||||
|
|
||||||
@config_packages
|
@config_packages
|
||||||
def _deprecated_preferences(error_cls):
|
def _deprecated_preferences(error_cls):
|
||||||
"""Search package preferences deprecated in v0.21 (and slated for removal in v0.22)"""
|
"""Search package preferences deprecated in v0.21 (and slated for removal in v0.23)"""
|
||||||
# TODO (v0.22): remove this audit as the attributes will not be allowed in config
|
# TODO (v0.23): remove this audit as the attributes will not be allowed in config
|
||||||
errors = []
|
errors = []
|
||||||
packages_yaml = spack.config.CONFIG.get_config("packages")
|
packages_yaml = spack.config.CONFIG.get_config("packages")
|
||||||
|
|
||||||
@@ -421,6 +421,10 @@ def _check_patch_urls(pkgs, error_cls):
|
|||||||
r"^https?://(?:patch-diff\.)?github(?:usercontent)?\.com/"
|
r"^https?://(?:patch-diff\.)?github(?:usercontent)?\.com/"
|
||||||
r".+/.+/(?:commit|pull)/[a-fA-F0-9]+\.(?:patch|diff)"
|
r".+/.+/(?:commit|pull)/[a-fA-F0-9]+\.(?:patch|diff)"
|
||||||
)
|
)
|
||||||
|
github_pull_commits_re = (
|
||||||
|
r"^https?://(?:patch-diff\.)?github(?:usercontent)?\.com/"
|
||||||
|
r".+/.+/pull/\d+/commits/[a-fA-F0-9]+\.(?:patch|diff)"
|
||||||
|
)
|
||||||
# Only .diff URLs have stable/full hashes:
|
# Only .diff URLs have stable/full hashes:
|
||||||
# https://forum.gitlab.com/t/patches-with-full-index/29313
|
# https://forum.gitlab.com/t/patches-with-full-index/29313
|
||||||
gitlab_patch_url_re = (
|
gitlab_patch_url_re = (
|
||||||
@@ -436,14 +440,24 @@ def _check_patch_urls(pkgs, error_cls):
|
|||||||
if not isinstance(patch, spack.patch.UrlPatch):
|
if not isinstance(patch, spack.patch.UrlPatch):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
if re.match(github_patch_url_re, patch.url):
|
if re.match(github_pull_commits_re, patch.url):
|
||||||
|
url = re.sub(r"/pull/\d+/commits/", r"/commit/", patch.url)
|
||||||
|
url = re.sub(r"^(.*)(?<!full_index=1)$", r"\1?full_index=1", url)
|
||||||
|
errors.append(
|
||||||
|
error_cls(
|
||||||
|
f"patch URL in package {pkg_cls.name} "
|
||||||
|
+ "must not be a pull request commit; "
|
||||||
|
+ f"instead use {url}",
|
||||||
|
[patch.url],
|
||||||
|
)
|
||||||
|
)
|
||||||
|
elif re.match(github_patch_url_re, patch.url):
|
||||||
full_index_arg = "?full_index=1"
|
full_index_arg = "?full_index=1"
|
||||||
if not patch.url.endswith(full_index_arg):
|
if not patch.url.endswith(full_index_arg):
|
||||||
errors.append(
|
errors.append(
|
||||||
error_cls(
|
error_cls(
|
||||||
"patch URL in package {0} must end with {1}".format(
|
f"patch URL in package {pkg_cls.name} "
|
||||||
pkg_cls.name, full_index_arg
|
+ f"must end with {full_index_arg}",
|
||||||
),
|
|
||||||
[patch.url],
|
[patch.url],
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
@@ -451,9 +465,7 @@ def _check_patch_urls(pkgs, error_cls):
|
|||||||
if not patch.url.endswith(".diff"):
|
if not patch.url.endswith(".diff"):
|
||||||
errors.append(
|
errors.append(
|
||||||
error_cls(
|
error_cls(
|
||||||
"patch URL in package {0} must end with .diff".format(
|
f"patch URL in package {pkg_cls.name} must end with .diff",
|
||||||
pkg_cls.name
|
|
||||||
),
|
|
||||||
[patch.url],
|
[patch.url],
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
@@ -779,7 +791,7 @@ def check_virtual_with_variants(spec, msg):
|
|||||||
return
|
return
|
||||||
error = error_cls(
|
error = error_cls(
|
||||||
f"{pkg_name}: {msg}",
|
f"{pkg_name}: {msg}",
|
||||||
f"remove variants from '{spec}' in depends_on directive in {filename}",
|
[f"remove variants from '{spec}' in depends_on directive in {filename}"],
|
||||||
)
|
)
|
||||||
errors.append(error)
|
errors.append(error)
|
||||||
|
|
||||||
|
|||||||
@@ -29,6 +29,7 @@
|
|||||||
import llnl.util.lang
|
import llnl.util.lang
|
||||||
import llnl.util.tty as tty
|
import llnl.util.tty as tty
|
||||||
from llnl.util.filesystem import BaseDirectoryVisitor, mkdirp, visit_directory_tree
|
from llnl.util.filesystem import BaseDirectoryVisitor, mkdirp, visit_directory_tree
|
||||||
|
from llnl.util.symlink import readlink
|
||||||
|
|
||||||
import spack.caches
|
import spack.caches
|
||||||
import spack.cmd
|
import spack.cmd
|
||||||
@@ -658,7 +659,7 @@ def get_buildfile_manifest(spec):
|
|||||||
# 2. paths are used as strings.
|
# 2. paths are used as strings.
|
||||||
for rel_path in visitor.symlinks:
|
for rel_path in visitor.symlinks:
|
||||||
abs_path = os.path.join(root, rel_path)
|
abs_path = os.path.join(root, rel_path)
|
||||||
link = os.readlink(abs_path)
|
link = readlink(abs_path)
|
||||||
if os.path.isabs(link) and link.startswith(spack.store.STORE.layout.root):
|
if os.path.isabs(link) and link.startswith(spack.store.STORE.layout.root):
|
||||||
data["link_to_relocate"].append(rel_path)
|
data["link_to_relocate"].append(rel_path)
|
||||||
|
|
||||||
@@ -2001,6 +2002,7 @@ def install_root_node(spec, unsigned=False, force=False, sha256=None):
|
|||||||
with spack.util.path.filter_padding():
|
with spack.util.path.filter_padding():
|
||||||
tty.msg('Installing "{0}" from a buildcache'.format(spec.format()))
|
tty.msg('Installing "{0}" from a buildcache'.format(spec.format()))
|
||||||
extract_tarball(spec, download_result, force)
|
extract_tarball(spec, download_result, force)
|
||||||
|
spec.package.windows_establish_runtime_linkage()
|
||||||
spack.hooks.post_install(spec, False)
|
spack.hooks.post_install(spec, False)
|
||||||
spack.store.STORE.db.add(spec, spack.store.STORE.layout)
|
spack.store.STORE.db.add(spec, spack.store.STORE.layout)
|
||||||
|
|
||||||
|
|||||||
@@ -213,15 +213,18 @@ def _root_spec(spec_str: str) -> str:
|
|||||||
Args:
|
Args:
|
||||||
spec_str: spec to be bootstrapped. Must be without compiler and target.
|
spec_str: spec to be bootstrapped. Must be without compiler and target.
|
||||||
"""
|
"""
|
||||||
# Add a compiler requirement to the root spec.
|
# Add a compiler and platform requirement to the root spec.
|
||||||
platform = str(spack.platforms.host())
|
platform = str(spack.platforms.host())
|
||||||
|
|
||||||
if platform == "darwin":
|
if platform == "darwin":
|
||||||
spec_str += " %apple-clang"
|
spec_str += " %apple-clang"
|
||||||
|
elif platform == "windows":
|
||||||
|
spec_str += " %msvc"
|
||||||
elif platform == "linux":
|
elif platform == "linux":
|
||||||
spec_str += " %gcc"
|
spec_str += " %gcc"
|
||||||
elif platform == "freebsd":
|
elif platform == "freebsd":
|
||||||
spec_str += " %clang"
|
spec_str += " %clang"
|
||||||
|
spec_str += f" platform={platform}"
|
||||||
target = archspec.cpu.host().family
|
target = archspec.cpu.host().family
|
||||||
spec_str += f" target={target}"
|
spec_str += f" target={target}"
|
||||||
|
|
||||||
|
|||||||
@@ -129,10 +129,10 @@ def _bootstrap_config_scopes() -> Sequence["spack.config.ConfigScope"]:
|
|||||||
configuration_paths = (spack.config.CONFIGURATION_DEFAULTS_PATH, ("bootstrap", _config_path()))
|
configuration_paths = (spack.config.CONFIGURATION_DEFAULTS_PATH, ("bootstrap", _config_path()))
|
||||||
for name, path in configuration_paths:
|
for name, path in configuration_paths:
|
||||||
platform = spack.platforms.host().name
|
platform = spack.platforms.host().name
|
||||||
platform_scope = spack.config.ConfigScope(
|
platform_scope = spack.config.DirectoryConfigScope(
|
||||||
"/".join([name, platform]), os.path.join(path, platform)
|
f"{name}/{platform}", os.path.join(path, platform)
|
||||||
)
|
)
|
||||||
generic_scope = spack.config.ConfigScope(name, path)
|
generic_scope = spack.config.DirectoryConfigScope(name, path)
|
||||||
config_scopes.extend([generic_scope, platform_scope])
|
config_scopes.extend([generic_scope, platform_scope])
|
||||||
msg = "[BOOTSTRAP CONFIG SCOPE] name={0}, path={1}"
|
msg = "[BOOTSTRAP CONFIG SCOPE] name={0}, path={1}"
|
||||||
tty.debug(msg.format(generic_scope.name, generic_scope.path))
|
tty.debug(msg.format(generic_scope.name, generic_scope.path))
|
||||||
|
|||||||
@@ -43,7 +43,7 @@
|
|||||||
from collections import defaultdict
|
from collections import defaultdict
|
||||||
from enum import Flag, auto
|
from enum import Flag, auto
|
||||||
from itertools import chain
|
from itertools import chain
|
||||||
from typing import List, Set, Tuple
|
from typing import Dict, List, Set, Tuple
|
||||||
|
|
||||||
import llnl.util.tty as tty
|
import llnl.util.tty as tty
|
||||||
from llnl.string import plural
|
from llnl.string import plural
|
||||||
@@ -72,6 +72,7 @@
|
|||||||
import spack.store
|
import spack.store
|
||||||
import spack.subprocess_context
|
import spack.subprocess_context
|
||||||
import spack.user_environment
|
import spack.user_environment
|
||||||
|
import spack.util.executable
|
||||||
import spack.util.path
|
import spack.util.path
|
||||||
import spack.util.pattern
|
import spack.util.pattern
|
||||||
from spack import traverse
|
from spack import traverse
|
||||||
@@ -91,7 +92,7 @@
|
|||||||
)
|
)
|
||||||
from spack.util.executable import Executable
|
from spack.util.executable import Executable
|
||||||
from spack.util.log_parse import make_log_context, parse_log_events
|
from spack.util.log_parse import make_log_context, parse_log_events
|
||||||
from spack.util.module_cmd import load_module, module, path_from_modules
|
from spack.util.module_cmd import load_module, path_from_modules
|
||||||
|
|
||||||
#
|
#
|
||||||
# This can be set by the user to globally disable parallel builds.
|
# This can be set by the user to globally disable parallel builds.
|
||||||
@@ -190,14 +191,6 @@ def __call__(self, *args, **kwargs):
|
|||||||
return super().__call__(*args, **kwargs)
|
return super().__call__(*args, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
def _on_cray():
|
|
||||||
host_platform = spack.platforms.host()
|
|
||||||
host_os = host_platform.operating_system("default_os")
|
|
||||||
on_cray = str(host_platform) == "cray"
|
|
||||||
using_cnl = re.match(r"cnl\d+", str(host_os))
|
|
||||||
return on_cray, using_cnl
|
|
||||||
|
|
||||||
|
|
||||||
def clean_environment():
|
def clean_environment():
|
||||||
# Stuff in here sanitizes the build environment to eliminate
|
# Stuff in here sanitizes the build environment to eliminate
|
||||||
# anything the user has set that may interfere. We apply it immediately
|
# anything the user has set that may interfere. We apply it immediately
|
||||||
@@ -241,17 +234,6 @@ def clean_environment():
|
|||||||
if varname.endswith("_ROOT") and varname != "SPACK_ROOT":
|
if varname.endswith("_ROOT") and varname != "SPACK_ROOT":
|
||||||
env.unset(varname)
|
env.unset(varname)
|
||||||
|
|
||||||
# On Cray "cluster" systems, unset CRAY_LD_LIBRARY_PATH to avoid
|
|
||||||
# interference with Spack dependencies.
|
|
||||||
# CNL requires these variables to be set (or at least some of them,
|
|
||||||
# depending on the CNL version).
|
|
||||||
on_cray, using_cnl = _on_cray()
|
|
||||||
if on_cray and not using_cnl:
|
|
||||||
env.unset("CRAY_LD_LIBRARY_PATH")
|
|
||||||
for varname in os.environ.keys():
|
|
||||||
if "PKGCONF" in varname:
|
|
||||||
env.unset(varname)
|
|
||||||
|
|
||||||
# Unset the following variables because they can affect installation of
|
# Unset the following variables because they can affect installation of
|
||||||
# Autotools and CMake packages.
|
# Autotools and CMake packages.
|
||||||
build_system_vars = [
|
build_system_vars = [
|
||||||
@@ -381,11 +363,7 @@ def set_compiler_environment_variables(pkg, env):
|
|||||||
_add_werror_handling(keep_werror, env)
|
_add_werror_handling(keep_werror, env)
|
||||||
|
|
||||||
# Set the target parameters that the compiler will add
|
# Set the target parameters that the compiler will add
|
||||||
# Don't set on cray platform because the targeting module handles this
|
isa_arg = spec.architecture.target.optimization_flags(compiler)
|
||||||
if spec.satisfies("platform=cray"):
|
|
||||||
isa_arg = ""
|
|
||||||
else:
|
|
||||||
isa_arg = spec.architecture.target.optimization_flags(compiler)
|
|
||||||
env.set("SPACK_TARGET_ARGS", isa_arg)
|
env.set("SPACK_TARGET_ARGS", isa_arg)
|
||||||
|
|
||||||
# Trap spack-tracked compiler flags as appropriate.
|
# Trap spack-tracked compiler flags as appropriate.
|
||||||
@@ -481,10 +459,7 @@ def set_wrapper_variables(pkg, env):
|
|||||||
|
|
||||||
# Find ccache binary and hand it to build environment
|
# Find ccache binary and hand it to build environment
|
||||||
if spack.config.get("config:ccache"):
|
if spack.config.get("config:ccache"):
|
||||||
ccache = Executable("ccache")
|
env.set(SPACK_CCACHE_BINARY, spack.util.executable.which_string("ccache", required=True))
|
||||||
if not ccache:
|
|
||||||
raise RuntimeError("No ccache binary found in PATH")
|
|
||||||
env.set(SPACK_CCACHE_BINARY, ccache)
|
|
||||||
|
|
||||||
# Gather information about various types of dependencies
|
# Gather information about various types of dependencies
|
||||||
link_deps = set(pkg.spec.traverse(root=False, deptype=("link")))
|
link_deps = set(pkg.spec.traverse(root=False, deptype=("link")))
|
||||||
@@ -730,12 +705,28 @@ def _static_to_shared_library(arch, compiler, static_lib, shared_lib=None, **kwa
|
|||||||
return compiler(*compiler_args, output=compiler_output)
|
return compiler(*compiler_args, output=compiler_output)
|
||||||
|
|
||||||
|
|
||||||
def get_rpath_deps(pkg):
|
def _get_rpath_deps_from_spec(
|
||||||
"""Return immediate or transitive RPATHs depending on the package."""
|
spec: spack.spec.Spec, transitive_rpaths: bool
|
||||||
if pkg.transitive_rpaths:
|
) -> List[spack.spec.Spec]:
|
||||||
return [d for d in pkg.spec.traverse(root=False, deptype=("link"))]
|
if not transitive_rpaths:
|
||||||
else:
|
return spec.dependencies(deptype=dt.LINK)
|
||||||
return pkg.spec.dependencies(deptype="link")
|
|
||||||
|
by_name: Dict[str, spack.spec.Spec] = {}
|
||||||
|
|
||||||
|
for dep in spec.traverse(root=False, deptype=dt.LINK):
|
||||||
|
lookup = by_name.get(dep.name)
|
||||||
|
if lookup is None:
|
||||||
|
by_name[dep.name] = dep
|
||||||
|
elif lookup.version < dep.version:
|
||||||
|
by_name[dep.name] = dep
|
||||||
|
|
||||||
|
return list(by_name.values())
|
||||||
|
|
||||||
|
|
||||||
|
def get_rpath_deps(pkg: spack.package_base.PackageBase) -> List[spack.spec.Spec]:
|
||||||
|
"""Return immediate or transitive dependencies (depending on the package) that need to be
|
||||||
|
rpath'ed. If a package occurs multiple times, the newest version is kept."""
|
||||||
|
return _get_rpath_deps_from_spec(pkg.spec, pkg.transitive_rpaths)
|
||||||
|
|
||||||
|
|
||||||
def get_rpaths(pkg):
|
def get_rpaths(pkg):
|
||||||
@@ -747,7 +738,9 @@ def get_rpaths(pkg):
|
|||||||
# Second module is our compiler mod name. We use that to get rpaths from
|
# Second module is our compiler mod name. We use that to get rpaths from
|
||||||
# module show output.
|
# module show output.
|
||||||
if pkg.compiler.modules and len(pkg.compiler.modules) > 1:
|
if pkg.compiler.modules and len(pkg.compiler.modules) > 1:
|
||||||
rpaths.append(path_from_modules([pkg.compiler.modules[1]]))
|
mod_rpath = path_from_modules([pkg.compiler.modules[1]])
|
||||||
|
if mod_rpath:
|
||||||
|
rpaths.append(mod_rpath)
|
||||||
return list(dedupe(filter_system_paths(rpaths)))
|
return list(dedupe(filter_system_paths(rpaths)))
|
||||||
|
|
||||||
|
|
||||||
@@ -817,14 +810,6 @@ def setup_package(pkg, dirty, context: Context = Context.BUILD):
|
|||||||
for mod in pkg.compiler.modules:
|
for mod in pkg.compiler.modules:
|
||||||
load_module(mod)
|
load_module(mod)
|
||||||
|
|
||||||
# kludge to handle cray mpich and libsci being automatically loaded by
|
|
||||||
# PrgEnv modules on cray platform. Module unload does no damage when
|
|
||||||
# unnecessary
|
|
||||||
on_cray, _ = _on_cray()
|
|
||||||
if on_cray and not dirty:
|
|
||||||
for mod in ["cray-mpich", "cray-libsci"]:
|
|
||||||
module("unload", mod)
|
|
||||||
|
|
||||||
if target and target.module_name:
|
if target and target.module_name:
|
||||||
load_module(target.module_name)
|
load_module(target.module_name)
|
||||||
|
|
||||||
@@ -1488,7 +1473,7 @@ def long_message(self):
|
|||||||
out.write(" {0}\n".format(self.log_name))
|
out.write(" {0}\n".format(self.log_name))
|
||||||
|
|
||||||
# Also output the test log path IF it exists
|
# Also output the test log path IF it exists
|
||||||
if self.context != "test":
|
if self.context != "test" and have_log:
|
||||||
test_log = join_path(os.path.dirname(self.log_name), spack_install_test_log)
|
test_log = join_path(os.path.dirname(self.log_name), spack_install_test_log)
|
||||||
if os.path.isfile(test_log):
|
if os.path.isfile(test_log):
|
||||||
out.write("\nSee test log for details:\n")
|
out.write("\nSee test log for details:\n")
|
||||||
|
|||||||
@@ -162,7 +162,9 @@ def initconfig_compiler_entries(self):
|
|||||||
ld_flags = " ".join(flags["ldflags"])
|
ld_flags = " ".join(flags["ldflags"])
|
||||||
ld_format_string = "CMAKE_{0}_LINKER_FLAGS"
|
ld_format_string = "CMAKE_{0}_LINKER_FLAGS"
|
||||||
# CMake has separate linker arguments for types of builds.
|
# CMake has separate linker arguments for types of builds.
|
||||||
for ld_type in ["EXE", "MODULE", "SHARED", "STATIC"]:
|
# 'ldflags' should not be used with CMAKE_STATIC_LINKER_FLAGS which
|
||||||
|
# is used by the archiver, so don't include "STATIC" in this loop:
|
||||||
|
for ld_type in ["EXE", "MODULE", "SHARED"]:
|
||||||
ld_string = ld_format_string.format(ld_type)
|
ld_string = ld_format_string.format(ld_type)
|
||||||
entries.append(cmake_cache_string(ld_string, ld_flags))
|
entries.append(cmake_cache_string(ld_string, ld_flags))
|
||||||
|
|
||||||
|
|||||||
@@ -110,9 +110,8 @@ def cuda_flags(arch_list):
|
|||||||
# From the NVIDIA install guide we know of conflicts for particular
|
# From the NVIDIA install guide we know of conflicts for particular
|
||||||
# platforms (linux, darwin), architectures (x86, powerpc) and compilers
|
# platforms (linux, darwin), architectures (x86, powerpc) and compilers
|
||||||
# (gcc, clang). We don't restrict %gcc and %clang conflicts to
|
# (gcc, clang). We don't restrict %gcc and %clang conflicts to
|
||||||
# platform=linux, since they should also apply to platform=cray, and may
|
# platform=linux, since they may apply to platform=darwin. We currently
|
||||||
# apply to platform=darwin. We currently do not provide conflicts for
|
# do not provide conflicts for platform=darwin with %apple-clang.
|
||||||
# platform=darwin with %apple-clang.
|
|
||||||
|
|
||||||
# Linux x86_64 compiler conflicts from here:
|
# Linux x86_64 compiler conflicts from here:
|
||||||
# https://gist.github.com/ax3l/9489132
|
# https://gist.github.com/ax3l/9489132
|
||||||
@@ -125,6 +124,8 @@ def cuda_flags(arch_list):
|
|||||||
# minimum supported versions
|
# minimum supported versions
|
||||||
conflicts("%gcc@:4", when="+cuda ^cuda@11.0:")
|
conflicts("%gcc@:4", when="+cuda ^cuda@11.0:")
|
||||||
conflicts("%gcc@:5", when="+cuda ^cuda@11.4:")
|
conflicts("%gcc@:5", when="+cuda ^cuda@11.4:")
|
||||||
|
conflicts("%gcc@:7.2", when="+cuda ^cuda@12.4:")
|
||||||
|
conflicts("%clang@:6", when="+cuda ^cuda@12.2:")
|
||||||
|
|
||||||
# maximum supported version
|
# maximum supported version
|
||||||
# NOTE:
|
# NOTE:
|
||||||
@@ -137,11 +138,14 @@ def cuda_flags(arch_list):
|
|||||||
conflicts("%gcc@11.2:", when="+cuda ^cuda@:11.5")
|
conflicts("%gcc@11.2:", when="+cuda ^cuda@:11.5")
|
||||||
conflicts("%gcc@12:", when="+cuda ^cuda@:11.8")
|
conflicts("%gcc@12:", when="+cuda ^cuda@:11.8")
|
||||||
conflicts("%gcc@13:", when="+cuda ^cuda@:12.3")
|
conflicts("%gcc@13:", when="+cuda ^cuda@:12.3")
|
||||||
|
conflicts("%gcc@14:", when="+cuda ^cuda@:12.5")
|
||||||
conflicts("%clang@12:", when="+cuda ^cuda@:11.4.0")
|
conflicts("%clang@12:", when="+cuda ^cuda@:11.4.0")
|
||||||
conflicts("%clang@13:", when="+cuda ^cuda@:11.5")
|
conflicts("%clang@13:", when="+cuda ^cuda@:11.5")
|
||||||
conflicts("%clang@14:", when="+cuda ^cuda@:11.7")
|
conflicts("%clang@14:", when="+cuda ^cuda@:11.7")
|
||||||
conflicts("%clang@15:", when="+cuda ^cuda@:12.0")
|
conflicts("%clang@15:", when="+cuda ^cuda@:12.0")
|
||||||
conflicts("%clang@16:", when="+cuda ^cuda@:12.3")
|
conflicts("%clang@16:", when="+cuda ^cuda@:12.1")
|
||||||
|
conflicts("%clang@17:", when="+cuda ^cuda@:12.3")
|
||||||
|
conflicts("%clang@18:", when="+cuda ^cuda@:12.5")
|
||||||
|
|
||||||
# https://gist.github.com/ax3l/9489132#gistcomment-3860114
|
# https://gist.github.com/ax3l/9489132#gistcomment-3860114
|
||||||
conflicts("%gcc@10", when="+cuda ^cuda@:11.4.0")
|
conflicts("%gcc@10", when="+cuda ^cuda@:11.4.0")
|
||||||
@@ -209,12 +213,16 @@ def cuda_flags(arch_list):
|
|||||||
conflicts("%intel@19.0:", when="+cuda ^cuda@:10.0")
|
conflicts("%intel@19.0:", when="+cuda ^cuda@:10.0")
|
||||||
conflicts("%intel@19.1:", when="+cuda ^cuda@:10.1")
|
conflicts("%intel@19.1:", when="+cuda ^cuda@:10.1")
|
||||||
conflicts("%intel@19.2:", when="+cuda ^cuda@:11.1.0")
|
conflicts("%intel@19.2:", when="+cuda ^cuda@:11.1.0")
|
||||||
|
conflicts("%intel@2021:", when="+cuda ^cuda@:11.4.0")
|
||||||
|
|
||||||
# XL is mostly relevant for ppc64le Linux
|
# XL is mostly relevant for ppc64le Linux
|
||||||
conflicts("%xl@:12,14:", when="+cuda ^cuda@:9.1")
|
conflicts("%xl@:12,14:", when="+cuda ^cuda@:9.1")
|
||||||
conflicts("%xl@:12,14:15,17:", when="+cuda ^cuda@9.2")
|
conflicts("%xl@:12,14:15,17:", when="+cuda ^cuda@9.2")
|
||||||
conflicts("%xl@:12,17:", when="+cuda ^cuda@:11.1.0")
|
conflicts("%xl@:12,17:", when="+cuda ^cuda@:11.1.0")
|
||||||
|
|
||||||
|
# PowerPC.
|
||||||
|
conflicts("target=ppc64le", when="+cuda ^cuda@12.5:")
|
||||||
|
|
||||||
# Darwin.
|
# Darwin.
|
||||||
# TODO: add missing conflicts for %apple-clang cuda@:10
|
# TODO: add missing conflicts for %apple-clang cuda@:10
|
||||||
conflicts("platform=darwin", when="+cuda ^cuda@11.0.2: ")
|
conflicts("platform=darwin", when="+cuda ^cuda@11.0.2:")
|
||||||
|
|||||||
@@ -72,7 +72,7 @@ def build_directory(self):
|
|||||||
def build_args(self):
|
def build_args(self):
|
||||||
"""Arguments for ``go build``."""
|
"""Arguments for ``go build``."""
|
||||||
# Pass ldflags -s = --strip-all and -w = --no-warnings by default
|
# Pass ldflags -s = --strip-all and -w = --no-warnings by default
|
||||||
return ["-ldflags", "-s -w", "-o", f"{self.pkg.name}"]
|
return ["-modcacherw", "-ldflags", "-s -w", "-o", f"{self.pkg.name}"]
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def check_args(self):
|
def check_args(self):
|
||||||
|
|||||||
@@ -846,6 +846,7 @@ def scalapack_libs(self):
|
|||||||
"^mpich@2:" in spec_root
|
"^mpich@2:" in spec_root
|
||||||
or "^cray-mpich" in spec_root
|
or "^cray-mpich" in spec_root
|
||||||
or "^mvapich2" in spec_root
|
or "^mvapich2" in spec_root
|
||||||
|
or "^mvapich" in spec_root
|
||||||
or "^intel-mpi" in spec_root
|
or "^intel-mpi" in spec_root
|
||||||
or "^intel-oneapi-mpi" in spec_root
|
or "^intel-oneapi-mpi" in spec_root
|
||||||
or "^intel-parallel-studio" in spec_root
|
or "^intel-parallel-studio" in spec_root
|
||||||
@@ -936,32 +937,15 @@ def mpi_setup_dependent_build_environment(self, env, dependent_spec, compilers_o
|
|||||||
"I_MPI_ROOT": self.normalize_path("mpi"),
|
"I_MPI_ROOT": self.normalize_path("mpi"),
|
||||||
}
|
}
|
||||||
|
|
||||||
# CAUTION - SIMILAR code in:
|
compiler_wrapper_commands = self.mpi_compiler_wrappers
|
||||||
# var/spack/repos/builtin/packages/mpich/package.py
|
wrapper_vars.update(
|
||||||
# var/spack/repos/builtin/packages/openmpi/package.py
|
{
|
||||||
# var/spack/repos/builtin/packages/mvapich2/package.py
|
"MPICC": compiler_wrapper_commands["MPICC"],
|
||||||
#
|
"MPICXX": compiler_wrapper_commands["MPICXX"],
|
||||||
# On Cray, the regular compiler wrappers *are* the MPI wrappers.
|
"MPIF77": compiler_wrapper_commands["MPIF77"],
|
||||||
if "platform=cray" in self.spec:
|
"MPIF90": compiler_wrapper_commands["MPIF90"],
|
||||||
# TODO: Confirm
|
}
|
||||||
wrapper_vars.update(
|
)
|
||||||
{
|
|
||||||
"MPICC": compilers_of_client["CC"],
|
|
||||||
"MPICXX": compilers_of_client["CXX"],
|
|
||||||
"MPIF77": compilers_of_client["F77"],
|
|
||||||
"MPIF90": compilers_of_client["F90"],
|
|
||||||
}
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
compiler_wrapper_commands = self.mpi_compiler_wrappers
|
|
||||||
wrapper_vars.update(
|
|
||||||
{
|
|
||||||
"MPICC": compiler_wrapper_commands["MPICC"],
|
|
||||||
"MPICXX": compiler_wrapper_commands["MPICXX"],
|
|
||||||
"MPIF77": compiler_wrapper_commands["MPIF77"],
|
|
||||||
"MPIF90": compiler_wrapper_commands["MPIF90"],
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
# Ensure that the directory containing the compiler wrappers is in the
|
# Ensure that the directory containing the compiler wrappers is in the
|
||||||
# PATH. Spack packages add `prefix.bin` to their dependents' paths,
|
# PATH. Spack packages add `prefix.bin` to their dependents' paths,
|
||||||
|
|||||||
@@ -24,7 +24,6 @@ class MSBuildPackage(spack.package_base.PackageBase):
|
|||||||
build_system("msbuild")
|
build_system("msbuild")
|
||||||
conflicts("platform=linux", when="build_system=msbuild")
|
conflicts("platform=linux", when="build_system=msbuild")
|
||||||
conflicts("platform=darwin", when="build_system=msbuild")
|
conflicts("platform=darwin", when="build_system=msbuild")
|
||||||
conflicts("platform=cray", when="build_system=msbuild")
|
|
||||||
|
|
||||||
|
|
||||||
@spack.builder.builder("msbuild")
|
@spack.builder.builder("msbuild")
|
||||||
|
|||||||
@@ -24,7 +24,6 @@ class NMakePackage(spack.package_base.PackageBase):
|
|||||||
build_system("nmake")
|
build_system("nmake")
|
||||||
conflicts("platform=linux", when="build_system=nmake")
|
conflicts("platform=linux", when="build_system=nmake")
|
||||||
conflicts("platform=darwin", when="build_system=nmake")
|
conflicts("platform=darwin", when="build_system=nmake")
|
||||||
conflicts("platform=cray", when="build_system=nmake")
|
|
||||||
|
|
||||||
|
|
||||||
@spack.builder.builder("nmake")
|
@spack.builder.builder("nmake")
|
||||||
@@ -145,7 +144,7 @@ def install(self, pkg, spec, prefix):
|
|||||||
opts += self.nmake_install_args()
|
opts += self.nmake_install_args()
|
||||||
if self.makefile_name:
|
if self.makefile_name:
|
||||||
opts.append("/F{}".format(self.makefile_name))
|
opts.append("/F{}".format(self.makefile_name))
|
||||||
opts.append(self.define("PREFIX", prefix))
|
opts.append(self.define("PREFIX", fs.windows_sfn(prefix)))
|
||||||
with fs.working_dir(self.build_directory):
|
with fs.working_dir(self.build_directory):
|
||||||
inspect.getmodule(self.pkg).nmake(
|
inspect.getmodule(self.pkg).nmake(
|
||||||
*opts, *self.install_targets, ignore_quotes=self.ignore_quotes
|
*opts, *self.install_targets, ignore_quotes=self.ignore_quotes
|
||||||
|
|||||||
@@ -36,9 +36,8 @@ class IntelOneApiPackage(Package):
|
|||||||
"target=ppc64:",
|
"target=ppc64:",
|
||||||
"target=ppc64le:",
|
"target=ppc64le:",
|
||||||
"target=aarch64:",
|
"target=aarch64:",
|
||||||
"platform=darwin:",
|
"platform=darwin",
|
||||||
"platform=cray:",
|
"platform=windows",
|
||||||
"platform=windows:",
|
|
||||||
]:
|
]:
|
||||||
conflicts(c, msg="This package in only available for x86_64 and Linux")
|
conflicts(c, msg="This package in only available for x86_64 and Linux")
|
||||||
|
|
||||||
|
|||||||
@@ -120,12 +120,6 @@ def skip_modules(self) -> Iterable[str]:
|
|||||||
"""
|
"""
|
||||||
return []
|
return []
|
||||||
|
|
||||||
@property
|
|
||||||
def python_spec(self):
|
|
||||||
"""Get python-venv if it exists or python otherwise."""
|
|
||||||
python, *_ = self.spec.dependencies("python-venv") or self.spec.dependencies("python")
|
|
||||||
return python
|
|
||||||
|
|
||||||
def view_file_conflicts(self, view, merge_map):
|
def view_file_conflicts(self, view, merge_map):
|
||||||
"""Report all file conflicts, excepting special cases for python.
|
"""Report all file conflicts, excepting special cases for python.
|
||||||
Specifically, this does not report errors for duplicate
|
Specifically, this does not report errors for duplicate
|
||||||
@@ -146,8 +140,12 @@ def view_file_conflicts(self, view, merge_map):
|
|||||||
def add_files_to_view(self, view, merge_map, skip_if_exists=True):
|
def add_files_to_view(self, view, merge_map, skip_if_exists=True):
|
||||||
# Patch up shebangs if the package extends Python and we put a Python interpreter in the
|
# Patch up shebangs if the package extends Python and we put a Python interpreter in the
|
||||||
# view.
|
# view.
|
||||||
python = self.python_spec
|
if not self.extendee_spec:
|
||||||
if not self.extendee_spec or python.external:
|
return super().add_files_to_view(view, merge_map, skip_if_exists)
|
||||||
|
|
||||||
|
python, *_ = self.spec.dependencies("python-venv") or self.spec.dependencies("python")
|
||||||
|
|
||||||
|
if python.external:
|
||||||
return super().add_files_to_view(view, merge_map, skip_if_exists)
|
return super().add_files_to_view(view, merge_map, skip_if_exists)
|
||||||
|
|
||||||
# We only patch shebangs in the bin directory.
|
# We only patch shebangs in the bin directory.
|
||||||
@@ -368,6 +366,12 @@ def list_url(cls) -> Optional[str]: # type: ignore[override]
|
|||||||
return f"https://pypi.org/simple/{name}/"
|
return f"https://pypi.org/simple/{name}/"
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
@property
|
||||||
|
def python_spec(self):
|
||||||
|
"""Get python-venv if it exists or python otherwise."""
|
||||||
|
python, *_ = self.spec.dependencies("python-venv") or self.spec.dependencies("python")
|
||||||
|
return python
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def headers(self) -> HeaderList:
|
def headers(self) -> HeaderList:
|
||||||
"""Discover header files in platlib."""
|
"""Discover header files in platlib."""
|
||||||
|
|||||||
@@ -34,6 +34,8 @@ def _misc_cache():
|
|||||||
return spack.util.file_cache.FileCache(path)
|
return spack.util.file_cache.FileCache(path)
|
||||||
|
|
||||||
|
|
||||||
|
FileCacheType = Union[spack.util.file_cache.FileCache, llnl.util.lang.Singleton]
|
||||||
|
|
||||||
#: Spack's cache for small data
|
#: Spack's cache for small data
|
||||||
MISC_CACHE: Union[spack.util.file_cache.FileCache, llnl.util.lang.Singleton] = (
|
MISC_CACHE: Union[spack.util.file_cache.FileCache, llnl.util.lang.Singleton] = (
|
||||||
llnl.util.lang.Singleton(_misc_cache)
|
llnl.util.lang.Singleton(_misc_cache)
|
||||||
|
|||||||
@@ -22,6 +22,8 @@
|
|||||||
from urllib.parse import urlencode
|
from urllib.parse import urlencode
|
||||||
from urllib.request import HTTPHandler, Request, build_opener
|
from urllib.request import HTTPHandler, Request, build_opener
|
||||||
|
|
||||||
|
import ruamel.yaml
|
||||||
|
|
||||||
import llnl.util.filesystem as fs
|
import llnl.util.filesystem as fs
|
||||||
import llnl.util.tty as tty
|
import llnl.util.tty as tty
|
||||||
from llnl.util.lang import memoized
|
from llnl.util.lang import memoized
|
||||||
@@ -44,6 +46,7 @@
|
|||||||
from spack import traverse
|
from spack import traverse
|
||||||
from spack.error import SpackError
|
from spack.error import SpackError
|
||||||
from spack.reporters import CDash, CDashConfiguration
|
from spack.reporters import CDash, CDashConfiguration
|
||||||
|
from spack.reporters.cdash import SPACK_CDASH_TIMEOUT
|
||||||
from spack.reporters.cdash import build_stamp as cdash_build_stamp
|
from spack.reporters.cdash import build_stamp as cdash_build_stamp
|
||||||
|
|
||||||
# See https://docs.gitlab.com/ee/ci/yaml/#retry for descriptions of conditions
|
# See https://docs.gitlab.com/ee/ci/yaml/#retry for descriptions of conditions
|
||||||
@@ -550,10 +553,9 @@ def generate_gitlab_ci_yaml(
|
|||||||
env,
|
env,
|
||||||
print_summary,
|
print_summary,
|
||||||
output_file,
|
output_file,
|
||||||
|
*,
|
||||||
prune_dag=False,
|
prune_dag=False,
|
||||||
check_index_only=False,
|
check_index_only=False,
|
||||||
run_optimizer=False,
|
|
||||||
use_dependencies=False,
|
|
||||||
artifacts_root=None,
|
artifacts_root=None,
|
||||||
remote_mirror_override=None,
|
remote_mirror_override=None,
|
||||||
):
|
):
|
||||||
@@ -574,12 +576,6 @@ def generate_gitlab_ci_yaml(
|
|||||||
this mode results in faster yaml generation time). Otherwise, also
|
this mode results in faster yaml generation time). Otherwise, also
|
||||||
check each spec directly by url (useful if there is no index or it
|
check each spec directly by url (useful if there is no index or it
|
||||||
might be out of date).
|
might be out of date).
|
||||||
run_optimizer (bool): If True, post-process the generated yaml to try
|
|
||||||
try to reduce the size (attempts to collect repeated configuration
|
|
||||||
and replace with definitions).)
|
|
||||||
use_dependencies (bool): If true, use "dependencies" rather than "needs"
|
|
||||||
("needs" allows DAG scheduling). Useful if gitlab instance cannot
|
|
||||||
be configured to handle more than a few "needs" per job.
|
|
||||||
artifacts_root (str): Path where artifacts like logs, environment
|
artifacts_root (str): Path where artifacts like logs, environment
|
||||||
files (spack.yaml, spack.lock), etc should be written. GitLab
|
files (spack.yaml, spack.lock), etc should be written. GitLab
|
||||||
requires this to be within the project directory.
|
requires this to be within the project directory.
|
||||||
@@ -683,6 +679,22 @@ def generate_gitlab_ci_yaml(
|
|||||||
"instead.",
|
"instead.",
|
||||||
)
|
)
|
||||||
|
|
||||||
|
def ensure_expected_target_path(path):
|
||||||
|
"""Returns passed paths with all Windows path separators exchanged
|
||||||
|
for posix separators only if copy_only_pipeline is enabled
|
||||||
|
|
||||||
|
This is required as copy_only_pipelines are a unique scenario where
|
||||||
|
the generate job and child pipelines are run on different platforms.
|
||||||
|
To make this compatible w/ Windows, we cannot write Windows style path separators
|
||||||
|
that will be consumed on by the Posix copy job runner.
|
||||||
|
|
||||||
|
TODO (johnwparent): Refactor config + cli read/write to deal only in posix
|
||||||
|
style paths
|
||||||
|
"""
|
||||||
|
if copy_only_pipeline and path:
|
||||||
|
path = path.replace("\\", "/")
|
||||||
|
return path
|
||||||
|
|
||||||
pipeline_mirrors = spack.mirror.MirrorCollection(binary=True)
|
pipeline_mirrors = spack.mirror.MirrorCollection(binary=True)
|
||||||
deprecated_mirror_config = False
|
deprecated_mirror_config = False
|
||||||
buildcache_destination = None
|
buildcache_destination = None
|
||||||
@@ -797,7 +809,8 @@ def generate_gitlab_ci_yaml(
|
|||||||
cli_scopes = [
|
cli_scopes = [
|
||||||
os.path.relpath(s.path, concrete_env_dir)
|
os.path.relpath(s.path, concrete_env_dir)
|
||||||
for s in cfg.scopes().values()
|
for s in cfg.scopes().values()
|
||||||
if isinstance(s, cfg.ImmutableConfigScope)
|
if not s.writable
|
||||||
|
and isinstance(s, (cfg.DirectoryConfigScope))
|
||||||
and s.path not in env_includes
|
and s.path not in env_includes
|
||||||
and os.path.exists(s.path)
|
and os.path.exists(s.path)
|
||||||
]
|
]
|
||||||
@@ -806,7 +819,7 @@ def generate_gitlab_ci_yaml(
|
|||||||
if scope not in include_scopes and scope not in env_includes:
|
if scope not in include_scopes and scope not in env_includes:
|
||||||
include_scopes.insert(0, scope)
|
include_scopes.insert(0, scope)
|
||||||
env_includes.extend(include_scopes)
|
env_includes.extend(include_scopes)
|
||||||
env_yaml_root["spack"]["include"] = env_includes
|
env_yaml_root["spack"]["include"] = [ensure_expected_target_path(i) for i in env_includes]
|
||||||
|
|
||||||
if "gitlab-ci" in env_yaml_root["spack"] and "ci" not in env_yaml_root["spack"]:
|
if "gitlab-ci" in env_yaml_root["spack"] and "ci" not in env_yaml_root["spack"]:
|
||||||
env_yaml_root["spack"]["ci"] = env_yaml_root["spack"].pop("gitlab-ci")
|
env_yaml_root["spack"]["ci"] = env_yaml_root["spack"].pop("gitlab-ci")
|
||||||
@@ -1227,6 +1240,9 @@ def main_script_replacements(cmd):
|
|||||||
"SPACK_REBUILD_EVERYTHING": str(rebuild_everything),
|
"SPACK_REBUILD_EVERYTHING": str(rebuild_everything),
|
||||||
"SPACK_REQUIRE_SIGNING": os.environ.get("SPACK_REQUIRE_SIGNING", "False"),
|
"SPACK_REQUIRE_SIGNING": os.environ.get("SPACK_REQUIRE_SIGNING", "False"),
|
||||||
}
|
}
|
||||||
|
output_vars = output_object["variables"]
|
||||||
|
for item, val in output_vars.items():
|
||||||
|
output_vars[item] = ensure_expected_target_path(val)
|
||||||
|
|
||||||
# TODO: Remove this block in Spack 0.23
|
# TODO: Remove this block in Spack 0.23
|
||||||
if deprecated_mirror_config and remote_mirror_override:
|
if deprecated_mirror_config and remote_mirror_override:
|
||||||
@@ -1251,17 +1267,6 @@ def main_script_replacements(cmd):
|
|||||||
with open(copy_specs_file, "w") as fd:
|
with open(copy_specs_file, "w") as fd:
|
||||||
fd.write(json.dumps(buildcache_copies))
|
fd.write(json.dumps(buildcache_copies))
|
||||||
|
|
||||||
# TODO(opadron): remove this or refactor
|
|
||||||
if run_optimizer:
|
|
||||||
import spack.ci_optimization as ci_opt
|
|
||||||
|
|
||||||
output_object = ci_opt.optimizer(output_object)
|
|
||||||
|
|
||||||
# TODO(opadron): remove this or refactor
|
|
||||||
if use_dependencies:
|
|
||||||
import spack.ci_needs_workaround as cinw
|
|
||||||
|
|
||||||
output_object = cinw.needs_to_dependencies(output_object)
|
|
||||||
else:
|
else:
|
||||||
# No jobs were generated
|
# No jobs were generated
|
||||||
noop_job = spack_ci_ir["jobs"]["noop"]["attributes"]
|
noop_job = spack_ci_ir["jobs"]["noop"]["attributes"]
|
||||||
@@ -1283,7 +1288,6 @@ def main_script_replacements(cmd):
|
|||||||
sorted_output = {}
|
sorted_output = {}
|
||||||
for output_key, output_value in sorted(output_object.items()):
|
for output_key, output_value in sorted(output_object.items()):
|
||||||
sorted_output[output_key] = output_value
|
sorted_output[output_key] = output_value
|
||||||
|
|
||||||
if known_broken_specs_encountered:
|
if known_broken_specs_encountered:
|
||||||
tty.error("This pipeline generated hashes known to be broken on develop:")
|
tty.error("This pipeline generated hashes known to be broken on develop:")
|
||||||
display_broken_spec_messages(broken_specs_url, known_broken_specs_encountered)
|
display_broken_spec_messages(broken_specs_url, known_broken_specs_encountered)
|
||||||
@@ -1291,8 +1295,11 @@ def main_script_replacements(cmd):
|
|||||||
if not rebuild_everything:
|
if not rebuild_everything:
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
with open(output_file, "w") as outf:
|
# Minimize yaml output size through use of anchors
|
||||||
outf.write(syaml.dump(sorted_output, default_flow_style=True))
|
syaml.anchorify(sorted_output)
|
||||||
|
|
||||||
|
with open(output_file, "w") as f:
|
||||||
|
ruamel.yaml.YAML().dump(sorted_output, f)
|
||||||
|
|
||||||
|
|
||||||
def _url_encode_string(input_string):
|
def _url_encode_string(input_string):
|
||||||
@@ -1478,6 +1485,12 @@ def copy_test_logs_to_artifacts(test_stage, job_test_dir):
|
|||||||
copy_files_to_artifacts(os.path.join(test_stage, "*", "*.txt"), job_test_dir)
|
copy_files_to_artifacts(os.path.join(test_stage, "*", "*.txt"), job_test_dir)
|
||||||
|
|
||||||
|
|
||||||
|
def win_quote(quote_str: str) -> str:
|
||||||
|
if IS_WINDOWS:
|
||||||
|
quote_str = f'"{quote_str}"'
|
||||||
|
return quote_str
|
||||||
|
|
||||||
|
|
||||||
def download_and_extract_artifacts(url, work_dir):
|
def download_and_extract_artifacts(url, work_dir):
|
||||||
"""Look for gitlab artifacts.zip at the given url, and attempt to download
|
"""Look for gitlab artifacts.zip at the given url, and attempt to download
|
||||||
and extract the contents into the given work_dir
|
and extract the contents into the given work_dir
|
||||||
@@ -1500,7 +1513,7 @@ def download_and_extract_artifacts(url, work_dir):
|
|||||||
request = Request(url, headers=headers)
|
request = Request(url, headers=headers)
|
||||||
request.get_method = lambda: "GET"
|
request.get_method = lambda: "GET"
|
||||||
|
|
||||||
response = opener.open(request)
|
response = opener.open(request, timeout=SPACK_CDASH_TIMEOUT)
|
||||||
response_code = response.getcode()
|
response_code = response.getcode()
|
||||||
|
|
||||||
if response_code != 200:
|
if response_code != 200:
|
||||||
@@ -1942,9 +1955,9 @@ def compose_command_err_handling(args):
|
|||||||
# but we need to handle EXEs (git, etc) ourselves
|
# but we need to handle EXEs (git, etc) ourselves
|
||||||
catch_exe_failure = (
|
catch_exe_failure = (
|
||||||
"""
|
"""
|
||||||
if ($LASTEXITCODE -ne 0){
|
if ($LASTEXITCODE -ne 0){{
|
||||||
throw "Command {} has failed"
|
throw 'Command {} has failed'
|
||||||
}
|
}}
|
||||||
"""
|
"""
|
||||||
if IS_WINDOWS
|
if IS_WINDOWS
|
||||||
else ""
|
else ""
|
||||||
@@ -2176,13 +2189,13 @@ def __init__(self, ci_cdash):
|
|||||||
def args(self):
|
def args(self):
|
||||||
return [
|
return [
|
||||||
"--cdash-upload-url",
|
"--cdash-upload-url",
|
||||||
self.upload_url,
|
win_quote(self.upload_url),
|
||||||
"--cdash-build",
|
"--cdash-build",
|
||||||
self.build_name,
|
win_quote(self.build_name),
|
||||||
"--cdash-site",
|
"--cdash-site",
|
||||||
self.site,
|
win_quote(self.site),
|
||||||
"--cdash-buildstamp",
|
"--cdash-buildstamp",
|
||||||
self.build_stamp,
|
win_quote(self.build_stamp),
|
||||||
]
|
]
|
||||||
|
|
||||||
@property # type: ignore
|
@property # type: ignore
|
||||||
@@ -2248,7 +2261,7 @@ def create_buildgroup(self, opener, headers, url, group_name, group_type):
|
|||||||
|
|
||||||
request = Request(url, data=enc_data, headers=headers)
|
request = Request(url, data=enc_data, headers=headers)
|
||||||
|
|
||||||
response = opener.open(request)
|
response = opener.open(request, timeout=SPACK_CDASH_TIMEOUT)
|
||||||
response_code = response.getcode()
|
response_code = response.getcode()
|
||||||
|
|
||||||
if response_code not in [200, 201]:
|
if response_code not in [200, 201]:
|
||||||
@@ -2294,7 +2307,7 @@ def populate_buildgroup(self, job_names):
|
|||||||
request = Request(url, data=enc_data, headers=headers)
|
request = Request(url, data=enc_data, headers=headers)
|
||||||
request.get_method = lambda: "PUT"
|
request.get_method = lambda: "PUT"
|
||||||
|
|
||||||
response = opener.open(request)
|
response = opener.open(request, timeout=SPACK_CDASH_TIMEOUT)
|
||||||
response_code = response.getcode()
|
response_code = response.getcode()
|
||||||
|
|
||||||
if response_code != 200:
|
if response_code != 200:
|
||||||
|
|||||||
@@ -1,34 +0,0 @@
|
|||||||
# Copyright 2013-2024 Lawrence Livermore National Security, LLC and other
|
|
||||||
# Spack Project Developers. See the top-level COPYRIGHT file for details.
|
|
||||||
#
|
|
||||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
|
||||||
import collections.abc
|
|
||||||
|
|
||||||
get_job_name = lambda needs_entry: (
|
|
||||||
needs_entry.get("job")
|
|
||||||
if (isinstance(needs_entry, collections.abc.Mapping) and needs_entry.get("artifacts", True))
|
|
||||||
else needs_entry if isinstance(needs_entry, str) else None
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def convert_job(job_entry):
|
|
||||||
if not isinstance(job_entry, collections.abc.Mapping):
|
|
||||||
return job_entry
|
|
||||||
|
|
||||||
needs = job_entry.get("needs")
|
|
||||||
if needs is None:
|
|
||||||
return job_entry
|
|
||||||
|
|
||||||
new_job = {}
|
|
||||||
new_job.update(job_entry)
|
|
||||||
del new_job["needs"]
|
|
||||||
|
|
||||||
new_job["dependencies"] = list(
|
|
||||||
filter((lambda x: x is not None), (get_job_name(needs_entry) for needs_entry in needs))
|
|
||||||
)
|
|
||||||
|
|
||||||
return new_job
|
|
||||||
|
|
||||||
|
|
||||||
def needs_to_dependencies(yaml):
|
|
||||||
return dict((k, convert_job(v)) for k, v in yaml.items())
|
|
||||||
@@ -1,363 +0,0 @@
|
|||||||
# Copyright 2013-2024 Lawrence Livermore National Security, LLC and other
|
|
||||||
# Spack Project Developers. See the top-level COPYRIGHT file for details.
|
|
||||||
#
|
|
||||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
|
||||||
import collections
|
|
||||||
import collections.abc
|
|
||||||
import copy
|
|
||||||
import hashlib
|
|
||||||
|
|
||||||
import spack.util.spack_yaml as syaml
|
|
||||||
|
|
||||||
|
|
||||||
def sort_yaml_obj(obj):
|
|
||||||
if isinstance(obj, collections.abc.Mapping):
|
|
||||||
return syaml.syaml_dict(
|
|
||||||
(k, sort_yaml_obj(v)) for k, v in sorted(obj.items(), key=(lambda item: str(item[0])))
|
|
||||||
)
|
|
||||||
|
|
||||||
if isinstance(obj, collections.abc.Sequence) and not isinstance(obj, str):
|
|
||||||
return syaml.syaml_list(sort_yaml_obj(x) for x in obj)
|
|
||||||
|
|
||||||
return obj
|
|
||||||
|
|
||||||
|
|
||||||
def matches(obj, proto):
|
|
||||||
"""Returns True if the test object "obj" matches the prototype object
|
|
||||||
"proto".
|
|
||||||
|
|
||||||
If obj and proto are mappings, obj matches proto if (key in obj) and
|
|
||||||
(obj[key] matches proto[key]) for every key in proto.
|
|
||||||
|
|
||||||
If obj and proto are sequences, obj matches proto if they are of the same
|
|
||||||
length and (a matches b) for every (a,b) in zip(obj, proto).
|
|
||||||
|
|
||||||
Otherwise, obj matches proto if obj == proto.
|
|
||||||
|
|
||||||
Precondition: proto must not have any reference cycles
|
|
||||||
"""
|
|
||||||
if isinstance(obj, collections.abc.Mapping):
|
|
||||||
if not isinstance(proto, collections.abc.Mapping):
|
|
||||||
return False
|
|
||||||
|
|
||||||
return all((key in obj and matches(obj[key], val)) for key, val in proto.items())
|
|
||||||
|
|
||||||
if isinstance(obj, collections.abc.Sequence) and not isinstance(obj, str):
|
|
||||||
if not (isinstance(proto, collections.abc.Sequence) and not isinstance(proto, str)):
|
|
||||||
return False
|
|
||||||
|
|
||||||
if len(obj) != len(proto):
|
|
||||||
return False
|
|
||||||
|
|
||||||
return all(matches(obj[index], val) for index, val in enumerate(proto))
|
|
||||||
|
|
||||||
return obj == proto
|
|
||||||
|
|
||||||
|
|
||||||
def subkeys(obj, proto):
|
|
||||||
"""Returns the test mapping "obj" after factoring out the items it has in
|
|
||||||
common with the prototype mapping "proto".
|
|
||||||
|
|
||||||
Consider a recursive merge operation, merge(a, b) on mappings a and b, that
|
|
||||||
returns a mapping, m, whose keys are the union of the keys of a and b, and
|
|
||||||
for every such key, "k", its corresponding value is:
|
|
||||||
|
|
||||||
- merge(a[key], b[key]) if a[key] and b[key] are mappings, or
|
|
||||||
- b[key] if (key in b) and not matches(a[key], b[key]),
|
|
||||||
or
|
|
||||||
- a[key] otherwise
|
|
||||||
|
|
||||||
|
|
||||||
If obj and proto are mappings, the returned object is the smallest object,
|
|
||||||
"a", such that merge(a, proto) matches obj.
|
|
||||||
|
|
||||||
Otherwise, obj is returned.
|
|
||||||
"""
|
|
||||||
if not (
|
|
||||||
isinstance(obj, collections.abc.Mapping) and isinstance(proto, collections.abc.Mapping)
|
|
||||||
):
|
|
||||||
return obj
|
|
||||||
|
|
||||||
new_obj = {}
|
|
||||||
for key, value in obj.items():
|
|
||||||
if key not in proto:
|
|
||||||
new_obj[key] = value
|
|
||||||
continue
|
|
||||||
|
|
||||||
if matches(value, proto[key]) and matches(proto[key], value):
|
|
||||||
continue
|
|
||||||
|
|
||||||
if isinstance(value, collections.abc.Mapping):
|
|
||||||
new_obj[key] = subkeys(value, proto[key])
|
|
||||||
continue
|
|
||||||
|
|
||||||
new_obj[key] = value
|
|
||||||
|
|
||||||
return new_obj
|
|
||||||
|
|
||||||
|
|
||||||
def add_extends(yaml, key):
|
|
||||||
"""Modifies the given object "yaml" so that it includes an "extends" key
|
|
||||||
whose value features "key".
|
|
||||||
|
|
||||||
If "extends" is not in yaml, then yaml is modified such that
|
|
||||||
yaml["extends"] == key.
|
|
||||||
|
|
||||||
If yaml["extends"] is a str, then yaml is modified such that
|
|
||||||
yaml["extends"] == [yaml["extends"], key]
|
|
||||||
|
|
||||||
If yaml["extends"] is a list that does not include key, then key is
|
|
||||||
appended to the list.
|
|
||||||
|
|
||||||
Otherwise, yaml is left unchanged.
|
|
||||||
"""
|
|
||||||
|
|
||||||
has_key = "extends" in yaml
|
|
||||||
extends = yaml.get("extends")
|
|
||||||
|
|
||||||
if has_key and not isinstance(extends, (str, collections.abc.Sequence)):
|
|
||||||
return
|
|
||||||
|
|
||||||
if extends is None:
|
|
||||||
yaml["extends"] = key
|
|
||||||
return
|
|
||||||
|
|
||||||
if isinstance(extends, str):
|
|
||||||
if extends != key:
|
|
||||||
yaml["extends"] = [extends, key]
|
|
||||||
return
|
|
||||||
|
|
||||||
if key not in extends:
|
|
||||||
extends.append(key)
|
|
||||||
|
|
||||||
|
|
||||||
def common_subobject(yaml, sub):
|
|
||||||
"""Factor prototype object "sub" out of the values of mapping "yaml".
|
|
||||||
|
|
||||||
Consider a modified copy of yaml, "new", where for each key, "key" in yaml:
|
|
||||||
|
|
||||||
- If yaml[key] matches sub, then new[key] = subkeys(yaml[key], sub).
|
|
||||||
- Otherwise, new[key] = yaml[key].
|
|
||||||
|
|
||||||
If the above match criteria is not satisfied for any such key, then (yaml,
|
|
||||||
None) is returned. The yaml object is returned unchanged.
|
|
||||||
|
|
||||||
Otherwise, each matching value in new is modified as in
|
|
||||||
add_extends(new[key], common_key), and then new[common_key] is set to sub.
|
|
||||||
The common_key value is chosen such that it does not match any preexisting
|
|
||||||
key in new. In this case, (new, common_key) is returned.
|
|
||||||
"""
|
|
||||||
match_list = set(k for k, v in yaml.items() if matches(v, sub))
|
|
||||||
|
|
||||||
if not match_list:
|
|
||||||
return yaml, None
|
|
||||||
|
|
||||||
common_prefix = ".c"
|
|
||||||
common_index = 0
|
|
||||||
|
|
||||||
while True:
|
|
||||||
common_key = "".join((common_prefix, str(common_index)))
|
|
||||||
if common_key not in yaml:
|
|
||||||
break
|
|
||||||
common_index += 1
|
|
||||||
|
|
||||||
new_yaml = {}
|
|
||||||
|
|
||||||
for key, val in yaml.items():
|
|
||||||
new_yaml[key] = copy.deepcopy(val)
|
|
||||||
|
|
||||||
if not matches(val, sub):
|
|
||||||
continue
|
|
||||||
|
|
||||||
new_yaml[key] = subkeys(new_yaml[key], sub)
|
|
||||||
add_extends(new_yaml[key], common_key)
|
|
||||||
|
|
||||||
new_yaml[common_key] = sub
|
|
||||||
|
|
||||||
return new_yaml, common_key
|
|
||||||
|
|
||||||
|
|
||||||
def print_delta(name, old, new, applied=None):
|
|
||||||
delta = new - old
|
|
||||||
reldelta = (1000 * delta) // old
|
|
||||||
reldelta = (reldelta // 10, reldelta % 10)
|
|
||||||
|
|
||||||
if applied is None:
|
|
||||||
applied = new <= old
|
|
||||||
|
|
||||||
print(
|
|
||||||
"\n".join(
|
|
||||||
(
|
|
||||||
"{0} {1}:",
|
|
||||||
" before: {2: 10d}",
|
|
||||||
" after : {3: 10d}",
|
|
||||||
" delta : {4:+10d} ({5:=+3d}.{6}%)",
|
|
||||||
)
|
|
||||||
).format(name, ("+" if applied else "x"), old, new, delta, reldelta[0], reldelta[1])
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def try_optimization_pass(name, yaml, optimization_pass, *args, **kwargs):
|
|
||||||
"""Try applying an optimization pass and return information about the
|
|
||||||
result
|
|
||||||
|
|
||||||
"name" is a string describing the nature of the pass. If it is a non-empty
|
|
||||||
string, summary statistics are also printed to stdout.
|
|
||||||
|
|
||||||
"yaml" is the object to apply the pass to.
|
|
||||||
|
|
||||||
"optimization_pass" is the function implementing the pass to be applied.
|
|
||||||
|
|
||||||
"args" and "kwargs" are the additional arguments to pass to optimization
|
|
||||||
pass. The pass is applied as
|
|
||||||
|
|
||||||
>>> (new_yaml, *other_results) = optimization_pass(yaml, *args, **kwargs)
|
|
||||||
|
|
||||||
The pass's results are greedily rejected if it does not modify the original
|
|
||||||
yaml document, or if it produces a yaml document that serializes to a
|
|
||||||
larger string.
|
|
||||||
|
|
||||||
Returns (new_yaml, yaml, applied, other_results) if applied, or
|
|
||||||
(yaml, new_yaml, applied, other_results) otherwise.
|
|
||||||
"""
|
|
||||||
result = optimization_pass(yaml, *args, **kwargs)
|
|
||||||
new_yaml, other_results = result[0], result[1:]
|
|
||||||
|
|
||||||
if new_yaml is yaml:
|
|
||||||
# pass was not applied
|
|
||||||
return (yaml, new_yaml, False, other_results)
|
|
||||||
|
|
||||||
pre_size = len(syaml.dump_config(sort_yaml_obj(yaml), default_flow_style=True))
|
|
||||||
post_size = len(syaml.dump_config(sort_yaml_obj(new_yaml), default_flow_style=True))
|
|
||||||
|
|
||||||
# pass makes the size worse: not applying
|
|
||||||
applied = post_size <= pre_size
|
|
||||||
if applied:
|
|
||||||
yaml, new_yaml = new_yaml, yaml
|
|
||||||
|
|
||||||
if name:
|
|
||||||
print_delta(name, pre_size, post_size, applied)
|
|
||||||
|
|
||||||
return (yaml, new_yaml, applied, other_results)
|
|
||||||
|
|
||||||
|
|
||||||
def build_histogram(iterator, key):
|
|
||||||
"""Builds a histogram of values given an iterable of mappings and a key.
|
|
||||||
|
|
||||||
For each mapping "m" with key "key" in iterator, the value m[key] is
|
|
||||||
considered.
|
|
||||||
|
|
||||||
Returns a list of tuples (hash, count, proportion, value), where
|
|
||||||
|
|
||||||
- "hash" is a sha1sum hash of the value.
|
|
||||||
- "count" is the number of occurences of values that hash to "hash".
|
|
||||||
- "proportion" is the proportion of all values considered above that
|
|
||||||
hash to "hash".
|
|
||||||
- "value" is one of the values considered above that hash to "hash".
|
|
||||||
Which value is chosen when multiple values hash to the same "hash" is
|
|
||||||
undefined.
|
|
||||||
|
|
||||||
The list is sorted in descending order by count, yielding the most
|
|
||||||
frequently occuring hashes first.
|
|
||||||
"""
|
|
||||||
buckets = collections.defaultdict(int)
|
|
||||||
values = {}
|
|
||||||
|
|
||||||
num_objects = 0
|
|
||||||
for obj in iterator:
|
|
||||||
num_objects += 1
|
|
||||||
|
|
||||||
try:
|
|
||||||
val = obj[key]
|
|
||||||
except (KeyError, TypeError):
|
|
||||||
continue
|
|
||||||
|
|
||||||
value_hash = hashlib.sha1()
|
|
||||||
value_hash.update(syaml.dump_config(sort_yaml_obj(val)).encode())
|
|
||||||
value_hash = value_hash.hexdigest()
|
|
||||||
|
|
||||||
buckets[value_hash] += 1
|
|
||||||
values[value_hash] = val
|
|
||||||
|
|
||||||
return [
|
|
||||||
(h, buckets[h], float(buckets[h]) / num_objects, values[h])
|
|
||||||
for h in sorted(buckets.keys(), key=lambda k: -buckets[k])
|
|
||||||
]
|
|
||||||
|
|
||||||
|
|
||||||
def optimizer(yaml):
|
|
||||||
original_size = len(syaml.dump_config(sort_yaml_obj(yaml), default_flow_style=True))
|
|
||||||
|
|
||||||
# try factoring out commonly repeated portions
|
|
||||||
common_job = {
|
|
||||||
"variables": {"SPACK_COMPILER_ACTION": "NONE"},
|
|
||||||
"after_script": ['rm -rf "./spack"'],
|
|
||||||
"artifacts": {"paths": ["jobs_scratch_dir", "cdash_report"], "when": "always"},
|
|
||||||
}
|
|
||||||
|
|
||||||
# look for a list of tags that appear frequently
|
|
||||||
_, count, proportion, tags = next(iter(build_histogram(yaml.values(), "tags")), (None,) * 4)
|
|
||||||
|
|
||||||
# If a list of tags is found, and there are more than one job that uses it,
|
|
||||||
# *and* the jobs that do use it represent at least 70% of all jobs, then
|
|
||||||
# add the list to the prototype object.
|
|
||||||
if tags and count > 1 and proportion >= 0.70:
|
|
||||||
common_job["tags"] = tags
|
|
||||||
|
|
||||||
# apply common object factorization
|
|
||||||
yaml, other, applied, rest = try_optimization_pass(
|
|
||||||
"general common object factorization", yaml, common_subobject, common_job
|
|
||||||
)
|
|
||||||
|
|
||||||
# look for a common script, and try factoring that out
|
|
||||||
_, count, proportion, script = next(
|
|
||||||
iter(build_histogram(yaml.values(), "script")), (None,) * 4
|
|
||||||
)
|
|
||||||
|
|
||||||
if script and count > 1 and proportion >= 0.70:
|
|
||||||
yaml, other, applied, rest = try_optimization_pass(
|
|
||||||
"script factorization", yaml, common_subobject, {"script": script}
|
|
||||||
)
|
|
||||||
|
|
||||||
# look for a common before_script, and try factoring that out
|
|
||||||
_, count, proportion, script = next(
|
|
||||||
iter(build_histogram(yaml.values(), "before_script")), (None,) * 4
|
|
||||||
)
|
|
||||||
|
|
||||||
if script and count > 1 and proportion >= 0.70:
|
|
||||||
yaml, other, applied, rest = try_optimization_pass(
|
|
||||||
"before_script factorization", yaml, common_subobject, {"before_script": script}
|
|
||||||
)
|
|
||||||
|
|
||||||
# Look specifically for the SPACK_ROOT_SPEC environment variables.
|
|
||||||
# Try to factor them out.
|
|
||||||
h = build_histogram(
|
|
||||||
(getattr(val, "get", lambda *args: {})("variables") for val in yaml.values()),
|
|
||||||
"SPACK_ROOT_SPEC",
|
|
||||||
)
|
|
||||||
|
|
||||||
# In this case, we try to factor out *all* instances of the SPACK_ROOT_SPEC
|
|
||||||
# environment variable; not just the one that appears with the greatest
|
|
||||||
# frequency. We only require that more than 1 job uses a given instance's
|
|
||||||
# value, because we expect the value to be very large, and so expect even
|
|
||||||
# few-to-one factorizations to yield large space savings.
|
|
||||||
counter = 0
|
|
||||||
for _, count, proportion, spec in h:
|
|
||||||
if count <= 1:
|
|
||||||
continue
|
|
||||||
|
|
||||||
counter += 1
|
|
||||||
|
|
||||||
yaml, other, applied, rest = try_optimization_pass(
|
|
||||||
"SPACK_ROOT_SPEC factorization ({count})".format(count=counter),
|
|
||||||
yaml,
|
|
||||||
common_subobject,
|
|
||||||
{"variables": {"SPACK_ROOT_SPEC": spec}},
|
|
||||||
)
|
|
||||||
|
|
||||||
new_size = len(syaml.dump_config(sort_yaml_obj(yaml), default_flow_style=True))
|
|
||||||
|
|
||||||
print("\n")
|
|
||||||
print_delta("overall summary", original_size, new_size)
|
|
||||||
print("\n")
|
|
||||||
return yaml
|
|
||||||
@@ -336,6 +336,7 @@ def display_specs(specs, args=None, **kwargs):
|
|||||||
groups (bool): display specs grouped by arch/compiler (default True)
|
groups (bool): display specs grouped by arch/compiler (default True)
|
||||||
decorator (typing.Callable): function to call to decorate specs
|
decorator (typing.Callable): function to call to decorate specs
|
||||||
all_headers (bool): show headers even when arch/compiler aren't defined
|
all_headers (bool): show headers even when arch/compiler aren't defined
|
||||||
|
status_fn (typing.Callable): if provided, prepend install-status info
|
||||||
output (typing.IO): A file object to write to. Default is ``sys.stdout``
|
output (typing.IO): A file object to write to. Default is ``sys.stdout``
|
||||||
|
|
||||||
"""
|
"""
|
||||||
@@ -359,6 +360,7 @@ def get_arg(name, default=None):
|
|||||||
groups = get_arg("groups", True)
|
groups = get_arg("groups", True)
|
||||||
all_headers = get_arg("all_headers", False)
|
all_headers = get_arg("all_headers", False)
|
||||||
output = get_arg("output", sys.stdout)
|
output = get_arg("output", sys.stdout)
|
||||||
|
status_fn = get_arg("status_fn", None)
|
||||||
|
|
||||||
decorator = get_arg("decorator", None)
|
decorator = get_arg("decorator", None)
|
||||||
if decorator is None:
|
if decorator is None:
|
||||||
@@ -386,6 +388,13 @@ def get_arg(name, default=None):
|
|||||||
def fmt(s, depth=0):
|
def fmt(s, depth=0):
|
||||||
"""Formatter function for all output specs"""
|
"""Formatter function for all output specs"""
|
||||||
string = ""
|
string = ""
|
||||||
|
|
||||||
|
if status_fn:
|
||||||
|
# This was copied from spec.tree's colorization logic
|
||||||
|
# then shortened because it seems like status_fn should
|
||||||
|
# always return an InstallStatus
|
||||||
|
string += colorize(status_fn(s).value)
|
||||||
|
|
||||||
if hashes:
|
if hashes:
|
||||||
string += gray_hash(s, hlen) + " "
|
string += gray_hash(s, hlen) + " "
|
||||||
string += depth * " "
|
string += depth * " "
|
||||||
@@ -444,7 +453,7 @@ def format_list(specs):
|
|||||||
def filter_loaded_specs(specs):
|
def filter_loaded_specs(specs):
|
||||||
"""Filter a list of specs returning only those that are
|
"""Filter a list of specs returning only those that are
|
||||||
currently loaded."""
|
currently loaded."""
|
||||||
hashes = os.environ.get(uenv.spack_loaded_hashes_var, "").split(":")
|
hashes = os.environ.get(uenv.spack_loaded_hashes_var, "").split(os.pathsep)
|
||||||
return [x for x in specs if x.dag_hash() in hashes]
|
return [x for x in specs if x.dag_hash() in hashes]
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -165,7 +165,7 @@ def _reset(args):
|
|||||||
if not ok_to_continue:
|
if not ok_to_continue:
|
||||||
raise RuntimeError("Aborting")
|
raise RuntimeError("Aborting")
|
||||||
|
|
||||||
for scope in spack.config.CONFIG.file_scopes:
|
for scope in spack.config.CONFIG.writable_scopes:
|
||||||
# The default scope should stay untouched
|
# The default scope should stay untouched
|
||||||
if scope.name == "defaults":
|
if scope.name == "defaults":
|
||||||
continue
|
continue
|
||||||
|
|||||||
@@ -13,7 +13,6 @@
|
|||||||
import shutil
|
import shutil
|
||||||
import sys
|
import sys
|
||||||
import tempfile
|
import tempfile
|
||||||
import urllib.request
|
|
||||||
from typing import Dict, List, Optional, Tuple, Union
|
from typing import Dict, List, Optional, Tuple, Union
|
||||||
|
|
||||||
import llnl.util.tty as tty
|
import llnl.util.tty as tty
|
||||||
@@ -54,6 +53,7 @@
|
|||||||
from spack.oci.oci import (
|
from spack.oci.oci import (
|
||||||
copy_missing_layers_with_retry,
|
copy_missing_layers_with_retry,
|
||||||
get_manifest_and_config_with_retry,
|
get_manifest_and_config_with_retry,
|
||||||
|
list_tags,
|
||||||
upload_blob_with_retry,
|
upload_blob_with_retry,
|
||||||
upload_manifest_with_retry,
|
upload_manifest_with_retry,
|
||||||
)
|
)
|
||||||
@@ -70,12 +70,6 @@ def setup_parser(subparser: argparse.ArgumentParser):
|
|||||||
|
|
||||||
push = subparsers.add_parser("push", aliases=["create"], help=push_fn.__doc__)
|
push = subparsers.add_parser("push", aliases=["create"], help=push_fn.__doc__)
|
||||||
push.add_argument("-f", "--force", action="store_true", help="overwrite tarball if it exists")
|
push.add_argument("-f", "--force", action="store_true", help="overwrite tarball if it exists")
|
||||||
push.add_argument(
|
|
||||||
"--allow-root",
|
|
||||||
"-a",
|
|
||||||
action="store_true",
|
|
||||||
help="allow install root string in binary files after RPATH substitution",
|
|
||||||
)
|
|
||||||
push_sign = push.add_mutually_exclusive_group(required=False)
|
push_sign = push.add_mutually_exclusive_group(required=False)
|
||||||
push_sign.add_argument(
|
push_sign.add_argument(
|
||||||
"--unsigned",
|
"--unsigned",
|
||||||
@@ -190,10 +184,6 @@ def setup_parser(subparser: argparse.ArgumentParser):
|
|||||||
keys.add_argument("-f", "--force", action="store_true", help="force new download of keys")
|
keys.add_argument("-f", "--force", action="store_true", help="force new download of keys")
|
||||||
keys.set_defaults(func=keys_fn)
|
keys.set_defaults(func=keys_fn)
|
||||||
|
|
||||||
preview = subparsers.add_parser("preview", help=preview_fn.__doc__)
|
|
||||||
arguments.add_common_arguments(preview, ["installed_specs"])
|
|
||||||
preview.set_defaults(func=preview_fn)
|
|
||||||
|
|
||||||
# Check if binaries need to be rebuilt on remote mirror
|
# Check if binaries need to be rebuilt on remote mirror
|
||||||
check = subparsers.add_parser("check", help=check_fn.__doc__)
|
check = subparsers.add_parser("check", help=check_fn.__doc__)
|
||||||
check.add_argument(
|
check.add_argument(
|
||||||
@@ -404,11 +394,6 @@ def push_fn(args):
|
|||||||
else:
|
else:
|
||||||
roots = spack.cmd.require_active_env(cmd_name="buildcache push").concrete_roots()
|
roots = spack.cmd.require_active_env(cmd_name="buildcache push").concrete_roots()
|
||||||
|
|
||||||
if args.allow_root:
|
|
||||||
tty.warn(
|
|
||||||
"The flag `--allow-root` is the default in Spack 0.21, will be removed in Spack 0.22"
|
|
||||||
)
|
|
||||||
|
|
||||||
mirror: spack.mirror.Mirror = args.mirror
|
mirror: spack.mirror.Mirror = args.mirror
|
||||||
|
|
||||||
# Check if this is an OCI image.
|
# Check if this is an OCI image.
|
||||||
@@ -856,10 +841,7 @@ def _config_from_tag(image_ref: ImageReference, tag: str) -> Optional[dict]:
|
|||||||
|
|
||||||
|
|
||||||
def _update_index_oci(image_ref: ImageReference, tmpdir: str, pool: MaybePool) -> None:
|
def _update_index_oci(image_ref: ImageReference, tmpdir: str, pool: MaybePool) -> None:
|
||||||
request = urllib.request.Request(url=image_ref.tags_url())
|
tags = list_tags(image_ref)
|
||||||
response = spack.oci.opener.urlopen(request)
|
|
||||||
spack.oci.opener.ensure_status(request, response, 200)
|
|
||||||
tags = json.load(response)["tags"]
|
|
||||||
|
|
||||||
# Fetch all image config files in parallel
|
# Fetch all image config files in parallel
|
||||||
spec_dicts = pool.starmap(
|
spec_dicts = pool.starmap(
|
||||||
@@ -963,14 +945,6 @@ def keys_fn(args):
|
|||||||
bindist.get_keys(args.install, args.trust, args.force)
|
bindist.get_keys(args.install, args.trust, args.force)
|
||||||
|
|
||||||
|
|
||||||
def preview_fn(args):
|
|
||||||
"""analyze an installed spec and reports whether executables and libraries are relocatable"""
|
|
||||||
tty.warn(
|
|
||||||
"`spack buildcache preview` is deprecated since `spack buildcache push --allow-root` is "
|
|
||||||
"now the default. This command will be removed in Spack 0.22"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def check_fn(args: argparse.Namespace):
|
def check_fn(args: argparse.Namespace):
|
||||||
"""check specs against remote binary mirror(s) to see if any need to be rebuilt
|
"""check specs against remote binary mirror(s) to see if any need to be rebuilt
|
||||||
|
|
||||||
|
|||||||
@@ -6,6 +6,7 @@
|
|||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
import shutil
|
import shutil
|
||||||
|
import warnings
|
||||||
from urllib.parse import urlparse, urlunparse
|
from urllib.parse import urlparse, urlunparse
|
||||||
|
|
||||||
import llnl.util.filesystem as fs
|
import llnl.util.filesystem as fs
|
||||||
@@ -31,7 +32,6 @@
|
|||||||
level = "long"
|
level = "long"
|
||||||
|
|
||||||
SPACK_COMMAND = "spack"
|
SPACK_COMMAND = "spack"
|
||||||
MAKE_COMMAND = "make"
|
|
||||||
INSTALL_FAIL_CODE = 1
|
INSTALL_FAIL_CODE = 1
|
||||||
FAILED_CREATE_BUILDCACHE_CODE = 100
|
FAILED_CREATE_BUILDCACHE_CODE = 100
|
||||||
|
|
||||||
@@ -40,6 +40,12 @@ def deindent(desc):
|
|||||||
return desc.replace(" ", "")
|
return desc.replace(" ", "")
|
||||||
|
|
||||||
|
|
||||||
|
def unicode_escape(path: str) -> str:
|
||||||
|
"""Returns transformed path with any unicode
|
||||||
|
characters replaced with their corresponding escapes"""
|
||||||
|
return path.encode("unicode-escape").decode("utf-8")
|
||||||
|
|
||||||
|
|
||||||
def setup_parser(subparser):
|
def setup_parser(subparser):
|
||||||
setup_parser.parser = subparser
|
setup_parser.parser = subparser
|
||||||
subparsers = subparser.add_subparsers(help="CI sub-commands")
|
subparsers = subparser.add_subparsers(help="CI sub-commands")
|
||||||
@@ -68,7 +74,7 @@ def setup_parser(subparser):
|
|||||||
"--optimize",
|
"--optimize",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
default=False,
|
default=False,
|
||||||
help="(experimental) optimize the gitlab yaml file for size\n\n"
|
help="(DEPRECATED) optimize the gitlab yaml file for size\n\n"
|
||||||
"run the generated document through a series of optimization passes "
|
"run the generated document through a series of optimization passes "
|
||||||
"designed to reduce the size of the generated file",
|
"designed to reduce the size of the generated file",
|
||||||
)
|
)
|
||||||
@@ -76,7 +82,7 @@ def setup_parser(subparser):
|
|||||||
"--dependencies",
|
"--dependencies",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
default=False,
|
default=False,
|
||||||
help="(experimental) disable DAG scheduling (use 'plain' dependencies)",
|
help="(DEPRECATED) disable DAG scheduling (use 'plain' dependencies)",
|
||||||
)
|
)
|
||||||
generate.add_argument(
|
generate.add_argument(
|
||||||
"--buildcache-destination",
|
"--buildcache-destination",
|
||||||
@@ -195,6 +201,18 @@ def ci_generate(args):
|
|||||||
before invoking this command. the value must be the CDash authorization token needed to create
|
before invoking this command. the value must be the CDash authorization token needed to create
|
||||||
a build group and register all generated jobs under it
|
a build group and register all generated jobs under it
|
||||||
"""
|
"""
|
||||||
|
if args.optimize:
|
||||||
|
warnings.warn(
|
||||||
|
"The --optimize option has been deprecated, and currently has no effect. "
|
||||||
|
"It will be removed in Spack v0.24."
|
||||||
|
)
|
||||||
|
|
||||||
|
if args.dependencies:
|
||||||
|
warnings.warn(
|
||||||
|
"The --dependencies option has been deprecated, and currently has no effect. "
|
||||||
|
"It will be removed in Spack v0.24."
|
||||||
|
)
|
||||||
|
|
||||||
env = spack.cmd.require_active_env(cmd_name="ci generate")
|
env = spack.cmd.require_active_env(cmd_name="ci generate")
|
||||||
|
|
||||||
if args.copy_to:
|
if args.copy_to:
|
||||||
@@ -207,8 +225,6 @@ def ci_generate(args):
|
|||||||
|
|
||||||
output_file = args.output_file
|
output_file = args.output_file
|
||||||
copy_yaml_to = args.copy_to
|
copy_yaml_to = args.copy_to
|
||||||
run_optimizer = args.optimize
|
|
||||||
use_dependencies = args.dependencies
|
|
||||||
prune_dag = args.prune_dag
|
prune_dag = args.prune_dag
|
||||||
index_only = args.index_only
|
index_only = args.index_only
|
||||||
artifacts_root = args.artifacts_root
|
artifacts_root = args.artifacts_root
|
||||||
@@ -229,8 +245,6 @@ def ci_generate(args):
|
|||||||
output_file,
|
output_file,
|
||||||
prune_dag=prune_dag,
|
prune_dag=prune_dag,
|
||||||
check_index_only=index_only,
|
check_index_only=index_only,
|
||||||
run_optimizer=run_optimizer,
|
|
||||||
use_dependencies=use_dependencies,
|
|
||||||
artifacts_root=artifacts_root,
|
artifacts_root=artifacts_root,
|
||||||
remote_mirror_override=buildcache_destination,
|
remote_mirror_override=buildcache_destination,
|
||||||
)
|
)
|
||||||
@@ -551,75 +565,35 @@ def ci_rebuild(args):
|
|||||||
# No hash match anywhere means we need to rebuild spec
|
# No hash match anywhere means we need to rebuild spec
|
||||||
|
|
||||||
# Start with spack arguments
|
# Start with spack arguments
|
||||||
spack_cmd = [SPACK_COMMAND, "--color=always", "--backtrace", "--verbose"]
|
spack_cmd = [SPACK_COMMAND, "--color=always", "--backtrace", "--verbose", "install"]
|
||||||
|
|
||||||
config = cfg.get("config")
|
config = cfg.get("config")
|
||||||
if not config["verify_ssl"]:
|
if not config["verify_ssl"]:
|
||||||
spack_cmd.append("-k")
|
spack_cmd.append("-k")
|
||||||
|
|
||||||
install_args = []
|
install_args = [f'--use-buildcache={spack_ci.win_quote("package:never,dependencies:only")}']
|
||||||
|
|
||||||
can_verify = spack_ci.can_verify_binaries()
|
can_verify = spack_ci.can_verify_binaries()
|
||||||
verify_binaries = can_verify and spack_is_pr_pipeline is False
|
verify_binaries = can_verify and spack_is_pr_pipeline is False
|
||||||
if not verify_binaries:
|
if not verify_binaries:
|
||||||
install_args.append("--no-check-signature")
|
install_args.append("--no-check-signature")
|
||||||
|
|
||||||
slash_hash = "/{}".format(job_spec.dag_hash())
|
slash_hash = spack_ci.win_quote("/" + job_spec.dag_hash())
|
||||||
|
|
||||||
# Arguments when installing dependencies from cache
|
|
||||||
deps_install_args = install_args
|
|
||||||
|
|
||||||
# Arguments when installing the root from sources
|
# Arguments when installing the root from sources
|
||||||
root_install_args = install_args + [
|
deps_install_args = install_args + ["--only=dependencies"]
|
||||||
"--keep-stage",
|
root_install_args = install_args + ["--keep-stage", "--only=package"]
|
||||||
"--only=package",
|
|
||||||
"--use-buildcache=package:never,dependencies:only",
|
|
||||||
]
|
|
||||||
if cdash_handler:
|
if cdash_handler:
|
||||||
# Add additional arguments to `spack install` for CDash reporting.
|
# Add additional arguments to `spack install` for CDash reporting.
|
||||||
root_install_args.extend(cdash_handler.args())
|
root_install_args.extend(cdash_handler.args())
|
||||||
root_install_args.append(slash_hash)
|
|
||||||
|
|
||||||
# ["x", "y"] -> "'x' 'y'"
|
|
||||||
args_to_string = lambda args: " ".join("'{}'".format(arg) for arg in args)
|
|
||||||
|
|
||||||
commands = [
|
commands = [
|
||||||
# apparently there's a race when spack bootstraps? do it up front once
|
# apparently there's a race when spack bootstraps? do it up front once
|
||||||
[SPACK_COMMAND, "-e", env.path, "bootstrap", "now"],
|
[SPACK_COMMAND, "-e", unicode_escape(env.path), "bootstrap", "now"],
|
||||||
[
|
spack_cmd + deps_install_args + [slash_hash],
|
||||||
SPACK_COMMAND,
|
spack_cmd + root_install_args + [slash_hash],
|
||||||
"-e",
|
|
||||||
env.path,
|
|
||||||
"env",
|
|
||||||
"depfile",
|
|
||||||
"-o",
|
|
||||||
"Makefile",
|
|
||||||
"--use-buildcache=package:never,dependencies:only",
|
|
||||||
slash_hash, # limit to spec we're building
|
|
||||||
],
|
|
||||||
[
|
|
||||||
# --output-sync requires GNU make 4.x.
|
|
||||||
# Old make errors when you pass it a flag it doesn't recognize,
|
|
||||||
# but it doesn't error or warn when you set unrecognized flags in
|
|
||||||
# this variable.
|
|
||||||
"export",
|
|
||||||
"GNUMAKEFLAGS=--output-sync=recurse",
|
|
||||||
],
|
|
||||||
[
|
|
||||||
MAKE_COMMAND,
|
|
||||||
"SPACK={}".format(args_to_string(spack_cmd)),
|
|
||||||
"SPACK_COLOR=always",
|
|
||||||
"SPACK_INSTALL_FLAGS={}".format(args_to_string(deps_install_args)),
|
|
||||||
"-j$(nproc)",
|
|
||||||
"install-deps/{}".format(
|
|
||||||
spack.environment.depfile.MakefileSpec(job_spec).safe_format(
|
|
||||||
"{name}-{version}-{hash}"
|
|
||||||
)
|
|
||||||
),
|
|
||||||
],
|
|
||||||
spack_cmd + ["install"] + root_install_args,
|
|
||||||
]
|
]
|
||||||
|
|
||||||
tty.debug("Installing {0} from source".format(job_spec.name))
|
tty.debug("Installing {0} from source".format(job_spec.name))
|
||||||
install_exit_code = spack_ci.process_command("install", commands, repro_dir)
|
install_exit_code = spack_ci.process_command("install", commands, repro_dir)
|
||||||
|
|
||||||
|
|||||||
@@ -106,7 +106,8 @@ def clean(parser, args):
|
|||||||
|
|
||||||
# Then do the cleaning falling through the cases
|
# Then do the cleaning falling through the cases
|
||||||
if args.specs:
|
if args.specs:
|
||||||
specs = spack.cmd.parse_specs(args.specs, concretize=True)
|
specs = spack.cmd.parse_specs(args.specs, concretize=False)
|
||||||
|
specs = list(spack.cmd.matching_spec_from_env(x) for x in specs)
|
||||||
for spec in specs:
|
for spec in specs:
|
||||||
msg = "Cleaning build stage [{0}]"
|
msg = "Cleaning build stage [{0}]"
|
||||||
tty.msg(msg.format(spec.short_spec))
|
tty.msg(msg.format(spec.short_spec))
|
||||||
|
|||||||
@@ -3,6 +3,9 @@
|
|||||||
#
|
#
|
||||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
||||||
|
|
||||||
|
import llnl.util.tty as tty
|
||||||
|
from llnl.string import plural
|
||||||
|
|
||||||
import spack.cmd
|
import spack.cmd
|
||||||
import spack.cmd.common.arguments
|
import spack.cmd.common.arguments
|
||||||
import spack.environment as ev
|
import spack.environment as ev
|
||||||
@@ -43,5 +46,9 @@ def concretize(parser, args):
|
|||||||
with env.write_transaction():
|
with env.write_transaction():
|
||||||
concretized_specs = env.concretize(force=args.force, tests=tests)
|
concretized_specs = env.concretize(force=args.force, tests=tests)
|
||||||
if not args.quiet:
|
if not args.quiet:
|
||||||
ev.display_specs(concretized_specs)
|
if concretized_specs:
|
||||||
|
tty.msg(f"Concretized {plural(len(concretized_specs), 'spec')}:")
|
||||||
|
ev.display_specs([concrete for _, concrete in concretized_specs])
|
||||||
|
else:
|
||||||
|
tty.msg("No new specs to concretize.")
|
||||||
env.write()
|
env.write()
|
||||||
|
|||||||
@@ -156,7 +156,7 @@ def print_flattened_configuration(*, blame: bool) -> None:
|
|||||||
"""
|
"""
|
||||||
env = ev.active_environment()
|
env = ev.active_environment()
|
||||||
if env is not None:
|
if env is not None:
|
||||||
pristine = env.manifest.pristine_yaml_content
|
pristine = env.manifest.yaml_content
|
||||||
flattened = pristine.copy()
|
flattened = pristine.copy()
|
||||||
flattened[spack.schema.env.TOP_LEVEL_KEY] = pristine[spack.schema.env.TOP_LEVEL_KEY].copy()
|
flattened[spack.schema.env.TOP_LEVEL_KEY] = pristine[spack.schema.env.TOP_LEVEL_KEY].copy()
|
||||||
else:
|
else:
|
||||||
@@ -264,7 +264,9 @@ def config_remove(args):
|
|||||||
def _can_update_config_file(scope: spack.config.ConfigScope, cfg_file):
|
def _can_update_config_file(scope: spack.config.ConfigScope, cfg_file):
|
||||||
if isinstance(scope, spack.config.SingleFileScope):
|
if isinstance(scope, spack.config.SingleFileScope):
|
||||||
return fs.can_access(cfg_file)
|
return fs.can_access(cfg_file)
|
||||||
return fs.can_write_to_dir(scope.path) and fs.can_access(cfg_file)
|
elif isinstance(scope, spack.config.DirectoryConfigScope):
|
||||||
|
return fs.can_write_to_dir(scope.path) and fs.can_access(cfg_file)
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
def _config_change_requires_scope(path, spec, scope, match_spec=None):
|
def _config_change_requires_scope(path, spec, scope, match_spec=None):
|
||||||
@@ -362,14 +364,11 @@ def config_change(args):
|
|||||||
def config_update(args):
|
def config_update(args):
|
||||||
# Read the configuration files
|
# Read the configuration files
|
||||||
spack.config.CONFIG.get_config(args.section, scope=args.scope)
|
spack.config.CONFIG.get_config(args.section, scope=args.scope)
|
||||||
updates: List[spack.config.ConfigScope] = list(
|
updates: List[spack.config.ConfigScope] = [
|
||||||
filter(
|
x
|
||||||
lambda s: not isinstance(
|
for x in spack.config.CONFIG.format_updates[args.section]
|
||||||
s, (spack.config.InternalConfigScope, spack.config.ImmutableConfigScope)
|
if not isinstance(x, spack.config.InternalConfigScope) and x.writable
|
||||||
),
|
]
|
||||||
spack.config.CONFIG.format_updates[args.section],
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
cannot_overwrite, skip_system_scope = [], False
|
cannot_overwrite, skip_system_scope = [], False
|
||||||
for scope in updates:
|
for scope in updates:
|
||||||
@@ -447,7 +446,7 @@ def _can_revert_update(scope_dir, cfg_file, bkp_file):
|
|||||||
|
|
||||||
|
|
||||||
def config_revert(args):
|
def config_revert(args):
|
||||||
scopes = [args.scope] if args.scope else [x.name for x in spack.config.CONFIG.file_scopes]
|
scopes = [args.scope] if args.scope else [x.name for x in spack.config.CONFIG.writable_scopes]
|
||||||
|
|
||||||
# Search for backup files in the configuration scopes
|
# Search for backup files in the configuration scopes
|
||||||
Entry = collections.namedtuple("Entry", ["scope", "cfg", "bkp"])
|
Entry = collections.namedtuple("Entry", ["scope", "cfg", "bkp"])
|
||||||
|
|||||||
@@ -2,7 +2,6 @@
|
|||||||
# Spack Project Developers. See the top-level COPYRIGHT file for details.
|
# Spack Project Developers. See the top-level COPYRIGHT file for details.
|
||||||
#
|
#
|
||||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import sys
|
import sys
|
||||||
@@ -934,7 +933,7 @@ def get_repository(args, name):
|
|||||||
# Figure out where the new package should live
|
# Figure out where the new package should live
|
||||||
repo_path = args.repo
|
repo_path = args.repo
|
||||||
if repo_path is not None:
|
if repo_path is not None:
|
||||||
repo = spack.repo.Repo(repo_path)
|
repo = spack.repo.from_path(repo_path)
|
||||||
if spec.namespace and spec.namespace != repo.namespace:
|
if spec.namespace and spec.namespace != repo.namespace:
|
||||||
tty.die(
|
tty.die(
|
||||||
"Can't create package with namespace {0} in repo with "
|
"Can't create package with namespace {0} in repo with "
|
||||||
@@ -942,9 +941,7 @@ def get_repository(args, name):
|
|||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
if spec.namespace:
|
if spec.namespace:
|
||||||
repo = spack.repo.PATH.get_repo(spec.namespace, None)
|
repo = spack.repo.PATH.get_repo(spec.namespace)
|
||||||
if not repo:
|
|
||||||
tty.die("Unknown namespace: '{0}'".format(spec.namespace))
|
|
||||||
else:
|
else:
|
||||||
repo = spack.repo.PATH.first_repo()
|
repo = spack.repo.PATH.first_repo()
|
||||||
|
|
||||||
|
|||||||
@@ -47,16 +47,6 @@ def inverted_dependencies():
|
|||||||
dependents of, e.g., `mpi`, but virtuals are not included as
|
dependents of, e.g., `mpi`, but virtuals are not included as
|
||||||
actual dependents.
|
actual dependents.
|
||||||
"""
|
"""
|
||||||
dag = {}
|
|
||||||
for pkg_cls in spack.repo.PATH.all_package_classes():
|
|
||||||
dag.setdefault(pkg_cls.name, set())
|
|
||||||
for dep in pkg_cls.dependencies_by_name():
|
|
||||||
deps = [dep]
|
|
||||||
|
|
||||||
# expand virtuals if necessary
|
|
||||||
if spack.repo.PATH.is_virtual(dep):
|
|
||||||
deps += [s.name for s in spack.repo.PATH.providers_for(dep)]
|
|
||||||
|
|
||||||
dag = collections.defaultdict(set)
|
dag = collections.defaultdict(set)
|
||||||
for pkg_cls in spack.repo.PATH.all_package_classes():
|
for pkg_cls in spack.repo.PATH.all_package_classes():
|
||||||
for _, deps_by_name in pkg_cls.dependencies.items():
|
for _, deps_by_name in pkg_cls.dependencies.items():
|
||||||
|
|||||||
@@ -9,6 +9,8 @@
|
|||||||
|
|
||||||
import spack.cmd
|
import spack.cmd
|
||||||
import spack.config
|
import spack.config
|
||||||
|
import spack.fetch_strategy
|
||||||
|
import spack.repo
|
||||||
import spack.spec
|
import spack.spec
|
||||||
import spack.util.path
|
import spack.util.path
|
||||||
import spack.version
|
import spack.version
|
||||||
@@ -69,13 +71,15 @@ def _retrieve_develop_source(spec, abspath):
|
|||||||
# We construct a package class ourselves, rather than asking for
|
# We construct a package class ourselves, rather than asking for
|
||||||
# Spec.package, since Spec only allows this when it is concrete
|
# Spec.package, since Spec only allows this when it is concrete
|
||||||
package = pkg_cls(spec)
|
package = pkg_cls(spec)
|
||||||
if isinstance(package.stage[0].fetcher, spack.fetch_strategy.GitFetchStrategy):
|
source_stage = package.stage[0]
|
||||||
package.stage[0].fetcher.get_full_repo = True
|
if isinstance(source_stage.fetcher, spack.fetch_strategy.GitFetchStrategy):
|
||||||
|
source_stage.fetcher.get_full_repo = True
|
||||||
# If we retrieved this version before and cached it, we may have
|
# If we retrieved this version before and cached it, we may have
|
||||||
# done so without cloning the full git repo; likewise, any
|
# done so without cloning the full git repo; likewise, any
|
||||||
# mirror might store an instance with truncated history.
|
# mirror might store an instance with truncated history.
|
||||||
package.stage[0].disable_mirrors()
|
source_stage.disable_mirrors()
|
||||||
|
|
||||||
|
source_stage.fetcher.set_package(package)
|
||||||
package.stage.steal_source(abspath)
|
package.stage.steal_source(abspath)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -3,6 +3,7 @@
|
|||||||
#
|
#
|
||||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
||||||
|
|
||||||
|
import errno
|
||||||
import glob
|
import glob
|
||||||
import os
|
import os
|
||||||
|
|
||||||
@@ -11,43 +12,13 @@
|
|||||||
import spack.cmd
|
import spack.cmd
|
||||||
import spack.paths
|
import spack.paths
|
||||||
import spack.repo
|
import spack.repo
|
||||||
from spack.spec import Spec
|
import spack.util.editor
|
||||||
from spack.util.editor import editor
|
|
||||||
|
|
||||||
description = "open package files in $EDITOR"
|
description = "open package files in $EDITOR"
|
||||||
section = "packaging"
|
section = "packaging"
|
||||||
level = "short"
|
level = "short"
|
||||||
|
|
||||||
|
|
||||||
def edit_package(name, repo_path, namespace):
|
|
||||||
"""Opens the requested package file in your favorite $EDITOR.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
name (str): The name of the package
|
|
||||||
repo_path (str): The path to the repository containing this package
|
|
||||||
namespace (str): A valid namespace registered with Spack
|
|
||||||
"""
|
|
||||||
# Find the location of the package
|
|
||||||
if repo_path:
|
|
||||||
repo = spack.repo.Repo(repo_path)
|
|
||||||
elif namespace:
|
|
||||||
repo = spack.repo.PATH.get_repo(namespace)
|
|
||||||
else:
|
|
||||||
repo = spack.repo.PATH
|
|
||||||
path = repo.filename_for_package_name(name)
|
|
||||||
|
|
||||||
spec = Spec(name)
|
|
||||||
if os.path.exists(path):
|
|
||||||
if not os.path.isfile(path):
|
|
||||||
tty.die("Something is wrong. '{0}' is not a file!".format(path))
|
|
||||||
if not os.access(path, os.R_OK):
|
|
||||||
tty.die("Insufficient permissions on '%s'!" % path)
|
|
||||||
else:
|
|
||||||
raise spack.repo.UnknownPackageError(spec.name)
|
|
||||||
|
|
||||||
editor(path)
|
|
||||||
|
|
||||||
|
|
||||||
def setup_parser(subparser):
|
def setup_parser(subparser):
|
||||||
excl_args = subparser.add_mutually_exclusive_group()
|
excl_args = subparser.add_mutually_exclusive_group()
|
||||||
|
|
||||||
@@ -98,41 +69,67 @@ def setup_parser(subparser):
|
|||||||
excl_args.add_argument("-r", "--repo", default=None, help="path to repo to edit package in")
|
excl_args.add_argument("-r", "--repo", default=None, help="path to repo to edit package in")
|
||||||
excl_args.add_argument("-N", "--namespace", default=None, help="namespace of package to edit")
|
excl_args.add_argument("-N", "--namespace", default=None, help="namespace of package to edit")
|
||||||
|
|
||||||
subparser.add_argument("package", nargs="?", default=None, help="package name")
|
subparser.add_argument("package", nargs="*", default=None, help="package name")
|
||||||
|
|
||||||
|
|
||||||
|
def locate_package(name: str, repo: spack.repo.Repo) -> str:
|
||||||
|
path = repo.filename_for_package_name(name)
|
||||||
|
|
||||||
|
try:
|
||||||
|
with open(path, "r"):
|
||||||
|
return path
|
||||||
|
except OSError as e:
|
||||||
|
if e.errno == errno.ENOENT:
|
||||||
|
raise spack.repo.UnknownPackageError(name) from e
|
||||||
|
tty.die(f"Cannot edit package: {e}")
|
||||||
|
|
||||||
|
|
||||||
|
def locate_file(name: str, path: str) -> str:
|
||||||
|
# convert command names to python module name
|
||||||
|
if path == spack.paths.command_path:
|
||||||
|
name = spack.cmd.python_name(name)
|
||||||
|
|
||||||
|
file_path = os.path.join(path, name)
|
||||||
|
|
||||||
|
# Try to open direct match.
|
||||||
|
try:
|
||||||
|
with open(file_path, "r"):
|
||||||
|
return file_path
|
||||||
|
except OSError as e:
|
||||||
|
if e.errno != errno.ENOENT:
|
||||||
|
tty.die(f"Cannot edit file: {e}")
|
||||||
|
pass
|
||||||
|
|
||||||
|
# Otherwise try to find a file that starts with the name
|
||||||
|
candidates = glob.glob(file_path + "*")
|
||||||
|
exclude_list = [".pyc", "~"] # exclude binaries and backups
|
||||||
|
files = [f for f in candidates if not any(f.endswith(ext) for ext in exclude_list)]
|
||||||
|
if len(files) > 1:
|
||||||
|
tty.die(
|
||||||
|
f"Multiple files start with `{name}`:\n"
|
||||||
|
+ "\n".join(f" {os.path.basename(f)}" for f in files)
|
||||||
|
)
|
||||||
|
elif not files:
|
||||||
|
tty.die(f"No file for '{name}' was found in {path}")
|
||||||
|
return files[0]
|
||||||
|
|
||||||
|
|
||||||
def edit(parser, args):
|
def edit(parser, args):
|
||||||
name = args.package
|
names = args.package
|
||||||
|
|
||||||
# By default, edit package files
|
|
||||||
path = spack.paths.packages_path
|
|
||||||
|
|
||||||
# If `--command`, `--test`, or `--module` is chosen, edit those instead
|
# If `--command`, `--test`, or `--module` is chosen, edit those instead
|
||||||
if args.path:
|
if args.path:
|
||||||
path = args.path
|
paths = [locate_file(name, args.path) for name in names] if names else [args.path]
|
||||||
if name:
|
spack.util.editor.editor(*paths)
|
||||||
# convert command names to python module name
|
elif names:
|
||||||
if path == spack.paths.command_path:
|
if args.repo:
|
||||||
name = spack.cmd.python_name(name)
|
repo = spack.repo.from_path(args.repo)
|
||||||
|
elif args.namespace:
|
||||||
path = os.path.join(path, name)
|
repo = spack.repo.PATH.get_repo(args.namespace)
|
||||||
if not os.path.exists(path):
|
else:
|
||||||
files = glob.glob(path + "*")
|
repo = spack.repo.PATH
|
||||||
exclude_list = [".pyc", "~"] # exclude binaries and backups
|
paths = [locate_package(name, repo) for name in names]
|
||||||
files = list(filter(lambda x: all(s not in x for s in exclude_list), files))
|
spack.util.editor.editor(*paths)
|
||||||
if len(files) > 1:
|
|
||||||
m = "Multiple files exist with the name {0}.".format(name)
|
|
||||||
m += " Please specify a suffix. Files are:\n\n"
|
|
||||||
for f in files:
|
|
||||||
m += " " + os.path.basename(f) + "\n"
|
|
||||||
tty.die(m)
|
|
||||||
if not files:
|
|
||||||
tty.die("No file for '{0}' was found in {1}".format(name, path))
|
|
||||||
path = files[0] # already confirmed only one entry in files
|
|
||||||
|
|
||||||
editor(path)
|
|
||||||
elif name:
|
|
||||||
edit_package(name, args.repo, args.namespace)
|
|
||||||
else:
|
else:
|
||||||
# By default open the directory where packages live
|
# By default open the directory where packages live
|
||||||
editor(path)
|
spack.util.editor.editor(spack.paths.packages_path)
|
||||||
|
|||||||
@@ -10,7 +10,7 @@
|
|||||||
import sys
|
import sys
|
||||||
import tempfile
|
import tempfile
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Optional
|
from typing import List, Optional
|
||||||
|
|
||||||
import llnl.string as string
|
import llnl.string as string
|
||||||
import llnl.util.filesystem as fs
|
import llnl.util.filesystem as fs
|
||||||
@@ -87,6 +87,9 @@ def env_create_setup_parser(subparser):
|
|||||||
default=None,
|
default=None,
|
||||||
help="either a lockfile (must end with '.json' or '.lock') or a manifest file",
|
help="either a lockfile (must end with '.json' or '.lock') or a manifest file",
|
||||||
)
|
)
|
||||||
|
subparser.add_argument(
|
||||||
|
"--include-concrete", action="append", help="name of old environment to copy specs from"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def env_create(args):
|
def env_create(args):
|
||||||
@@ -104,12 +107,17 @@ def env_create(args):
|
|||||||
# the environment should not include a view.
|
# the environment should not include a view.
|
||||||
with_view = None
|
with_view = None
|
||||||
|
|
||||||
|
include_concrete = None
|
||||||
|
if hasattr(args, "include_concrete"):
|
||||||
|
include_concrete = args.include_concrete
|
||||||
|
|
||||||
env = _env_create(
|
env = _env_create(
|
||||||
args.env_name,
|
args.env_name,
|
||||||
init_file=args.envfile,
|
init_file=args.envfile,
|
||||||
dir=args.dir or os.path.sep in args.env_name or args.env_name in (".", ".."),
|
dir=args.dir or os.path.sep in args.env_name or args.env_name in (".", ".."),
|
||||||
with_view=with_view,
|
with_view=with_view,
|
||||||
keep_relative=args.keep_relative,
|
keep_relative=args.keep_relative,
|
||||||
|
include_concrete=include_concrete,
|
||||||
)
|
)
|
||||||
|
|
||||||
# Generate views, only really useful for environments created from spack.lock files.
|
# Generate views, only really useful for environments created from spack.lock files.
|
||||||
@@ -123,31 +131,43 @@ def _env_create(
|
|||||||
dir: bool = False,
|
dir: bool = False,
|
||||||
with_view: Optional[str] = None,
|
with_view: Optional[str] = None,
|
||||||
keep_relative: bool = False,
|
keep_relative: bool = False,
|
||||||
|
include_concrete: Optional[List[str]] = None,
|
||||||
):
|
):
|
||||||
"""Create a new environment, with an optional yaml description.
|
"""Create a new environment, with an optional yaml description.
|
||||||
|
|
||||||
Arguments:
|
Arguments:
|
||||||
name_or_path: name of the environment to create, or path to it
|
name_or_path (str): name of the environment to create, or path to it
|
||||||
init_file: optional initialization file -- can be a JSON lockfile (*.lock, *.json) or YAML
|
init_file (str or file): optional initialization file -- can be
|
||||||
manifest file
|
a JSON lockfile (*.lock, *.json) or YAML manifest file
|
||||||
dir: if True, create an environment in a directory instead of a managed environment
|
dir (bool): if True, create an environment in a directory instead
|
||||||
keep_relative: if True, develop paths are copied verbatim into the new environment file,
|
of a named environment
|
||||||
otherwise they may be made absolute if the new environment is in a different location
|
keep_relative (bool): if True, develop paths are copied verbatim into
|
||||||
|
the new environment file, otherwise they may be made absolute if the
|
||||||
|
new environment is in a different location
|
||||||
|
include_concrete (list): list of the included concrete environments
|
||||||
"""
|
"""
|
||||||
if not dir:
|
if not dir:
|
||||||
env = ev.create(
|
env = ev.create(
|
||||||
name_or_path, init_file=init_file, with_view=with_view, keep_relative=keep_relative
|
name_or_path,
|
||||||
|
init_file=init_file,
|
||||||
|
with_view=with_view,
|
||||||
|
keep_relative=keep_relative,
|
||||||
|
include_concrete=include_concrete,
|
||||||
)
|
)
|
||||||
tty.msg(
|
tty.msg(
|
||||||
colorize(
|
colorize(
|
||||||
f"Created environment @c{{{cescape(env.name)}}} in: @c{{{cescape(env.path)}}}"
|
f"Created environment @c{{{cescape(name_or_path)}}} in: @c{{{cescape(env.path)}}}"
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
env = ev.create_in_dir(
|
env = ev.create_in_dir(
|
||||||
name_or_path, init_file=init_file, with_view=with_view, keep_relative=keep_relative
|
name_or_path,
|
||||||
|
init_file=init_file,
|
||||||
|
with_view=with_view,
|
||||||
|
keep_relative=keep_relative,
|
||||||
|
include_concrete=include_concrete,
|
||||||
)
|
)
|
||||||
tty.msg(colorize(f"Created anonymous environment in: @c{{{cescape(env.path)}}}"))
|
tty.msg(colorize(f"Created independent environment in: @c{{{cescape(env.path)}}}"))
|
||||||
tty.msg(f"Activate with: {colorize(f'@c{{spack env activate {cescape(name_or_path)}}}')}")
|
tty.msg(f"Activate with: {colorize(f'@c{{spack env activate {cescape(name_or_path)}}}')}")
|
||||||
return env
|
return env
|
||||||
|
|
||||||
@@ -434,6 +454,12 @@ def env_remove_setup_parser(subparser):
|
|||||||
"""remove an existing environment"""
|
"""remove an existing environment"""
|
||||||
subparser.add_argument("rm_env", metavar="env", nargs="+", help="environment(s) to remove")
|
subparser.add_argument("rm_env", metavar="env", nargs="+", help="environment(s) to remove")
|
||||||
arguments.add_common_arguments(subparser, ["yes_to_all"])
|
arguments.add_common_arguments(subparser, ["yes_to_all"])
|
||||||
|
subparser.add_argument(
|
||||||
|
"-f",
|
||||||
|
"--force",
|
||||||
|
action="store_true",
|
||||||
|
help="remove the environment even if it is included in another environment",
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def env_remove(args):
|
def env_remove(args):
|
||||||
@@ -443,13 +469,35 @@ def env_remove(args):
|
|||||||
and manifests embedded in repositories should be removed manually.
|
and manifests embedded in repositories should be removed manually.
|
||||||
"""
|
"""
|
||||||
read_envs = []
|
read_envs = []
|
||||||
|
valid_envs = []
|
||||||
bad_envs = []
|
bad_envs = []
|
||||||
for env_name in args.rm_env:
|
invalid_envs = []
|
||||||
|
|
||||||
|
for env_name in ev.all_environment_names():
|
||||||
try:
|
try:
|
||||||
env = ev.read(env_name)
|
env = ev.read(env_name)
|
||||||
read_envs.append(env)
|
valid_envs.append(env_name)
|
||||||
|
|
||||||
|
if env_name in args.rm_env:
|
||||||
|
read_envs.append(env)
|
||||||
except (spack.config.ConfigFormatError, ev.SpackEnvironmentConfigError):
|
except (spack.config.ConfigFormatError, ev.SpackEnvironmentConfigError):
|
||||||
bad_envs.append(env_name)
|
invalid_envs.append(env_name)
|
||||||
|
|
||||||
|
if env_name in args.rm_env:
|
||||||
|
bad_envs.append(env_name)
|
||||||
|
|
||||||
|
# Check if env is linked to another before trying to remove
|
||||||
|
for name in valid_envs:
|
||||||
|
# don't check if environment is included to itself
|
||||||
|
if name == env_name:
|
||||||
|
continue
|
||||||
|
environ = ev.Environment(ev.root(name))
|
||||||
|
if ev.root(env_name) in environ.included_concrete_envs:
|
||||||
|
msg = f'Environment "{env_name}" is being used by environment "{name}"'
|
||||||
|
if args.force:
|
||||||
|
tty.warn(msg)
|
||||||
|
else:
|
||||||
|
tty.die(msg)
|
||||||
|
|
||||||
if not args.yes_to_all:
|
if not args.yes_to_all:
|
||||||
environments = string.plural(len(args.rm_env), "environment", show_n=False)
|
environments = string.plural(len(args.rm_env), "environment", show_n=False)
|
||||||
|
|||||||
@@ -7,7 +7,7 @@
|
|||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import sys
|
import sys
|
||||||
from typing import List, Optional
|
from typing import List, Optional, Set
|
||||||
|
|
||||||
import llnl.util.tty as tty
|
import llnl.util.tty as tty
|
||||||
import llnl.util.tty.colify as colify
|
import llnl.util.tty.colify as colify
|
||||||
@@ -19,6 +19,7 @@
|
|||||||
import spack.detection
|
import spack.detection
|
||||||
import spack.error
|
import spack.error
|
||||||
import spack.repo
|
import spack.repo
|
||||||
|
import spack.spec
|
||||||
import spack.util.environment
|
import spack.util.environment
|
||||||
from spack.cmd.common import arguments
|
from spack.cmd.common import arguments
|
||||||
|
|
||||||
@@ -138,14 +139,26 @@ def external_find(args):
|
|||||||
candidate_packages, path_hints=args.path, max_workers=args.jobs
|
candidate_packages, path_hints=args.path, max_workers=args.jobs
|
||||||
)
|
)
|
||||||
|
|
||||||
new_entries = spack.detection.update_configuration(
|
new_specs = spack.detection.update_configuration(
|
||||||
detected_packages, scope=args.scope, buildable=not args.not_buildable
|
detected_packages, scope=args.scope, buildable=not args.not_buildable
|
||||||
)
|
)
|
||||||
if new_entries:
|
|
||||||
|
# If the user runs `spack external find --not-buildable mpich` we also mark `mpi` non-buildable
|
||||||
|
# to avoid that the concretizer picks a different mpi provider.
|
||||||
|
if new_specs and args.not_buildable:
|
||||||
|
virtuals: Set[str] = {
|
||||||
|
virtual.name
|
||||||
|
for new_spec in new_specs
|
||||||
|
for virtual_specs in spack.repo.PATH.get_pkg_class(new_spec.name).provided.values()
|
||||||
|
for virtual in virtual_specs
|
||||||
|
}
|
||||||
|
new_virtuals = spack.detection.set_virtuals_nonbuildable(virtuals, scope=args.scope)
|
||||||
|
new_specs.extend(spack.spec.Spec(name) for name in new_virtuals)
|
||||||
|
|
||||||
|
if new_specs:
|
||||||
path = spack.config.CONFIG.get_config_filename(args.scope, "packages")
|
path = spack.config.CONFIG.get_config_filename(args.scope, "packages")
|
||||||
msg = "The following specs have been detected on this system and added to {0}"
|
tty.msg(f"The following specs have been detected on this system and added to {path}")
|
||||||
tty.msg(msg.format(path))
|
spack.cmd.display_specs(new_specs)
|
||||||
spack.cmd.display_specs(new_entries)
|
|
||||||
else:
|
else:
|
||||||
tty.msg("No new external packages detected")
|
tty.msg("No new external packages detected")
|
||||||
|
|
||||||
|
|||||||
@@ -3,6 +3,7 @@
|
|||||||
#
|
#
|
||||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
||||||
|
|
||||||
|
import copy
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
import llnl.util.lang
|
import llnl.util.lang
|
||||||
@@ -45,6 +46,10 @@ def setup_parser(subparser):
|
|||||||
help="output specs as machine-readable json records",
|
help="output specs as machine-readable json records",
|
||||||
)
|
)
|
||||||
|
|
||||||
|
subparser.add_argument(
|
||||||
|
"-I", "--install-status", action="store_true", help="show install status of packages"
|
||||||
|
)
|
||||||
|
|
||||||
subparser.add_argument(
|
subparser.add_argument(
|
||||||
"-d", "--deps", action="store_true", help="output dependencies along with found specs"
|
"-d", "--deps", action="store_true", help="output dependencies along with found specs"
|
||||||
)
|
)
|
||||||
@@ -271,25 +276,45 @@ def root_decorator(spec, string):
|
|||||||
|
|
||||||
print()
|
print()
|
||||||
|
|
||||||
if args.show_concretized:
|
if env.included_concrete_envs:
|
||||||
tty.msg("Concretized roots")
|
tty.msg("Included specs")
|
||||||
cmd.display_specs(env.specs_by_hash.values(), args, decorator=decorator)
|
|
||||||
print()
|
|
||||||
|
|
||||||
# Display a header for the installed packages section IF there are installed
|
# Root specs cannot be displayed with prefixes, since those are not
|
||||||
# packages. If there aren't any, we'll just end up printing "0 installed packages"
|
# set for abstract specs. Same for hashes
|
||||||
# later.
|
root_args = copy.copy(args)
|
||||||
if results and not args.only_roots:
|
root_args.paths = False
|
||||||
tty.msg("Installed packages")
|
|
||||||
|
# Roots are displayed with variants, etc. so that we can see
|
||||||
|
# specifically what the user asked for.
|
||||||
|
cmd.display_specs(
|
||||||
|
env.included_user_specs,
|
||||||
|
root_args,
|
||||||
|
decorator=lambda s, f: color.colorize("@*{%s}" % f),
|
||||||
|
namespace=True,
|
||||||
|
show_flags=True,
|
||||||
|
show_full_compiler=True,
|
||||||
|
variants=True,
|
||||||
|
)
|
||||||
|
print()
|
||||||
|
|
||||||
|
|
||||||
def find(parser, args):
|
def find(parser, args):
|
||||||
q_args = query_arguments(args)
|
|
||||||
results = args.specs(**q_args)
|
|
||||||
|
|
||||||
env = ev.active_environment()
|
env = ev.active_environment()
|
||||||
|
|
||||||
if not env and args.only_roots:
|
if not env and args.only_roots:
|
||||||
tty.die("-r / --only-roots requires an active environment")
|
tty.die("-r / --only-roots requires an active environment")
|
||||||
|
if not env and args.show_concretized:
|
||||||
|
tty.die("-c / --show-concretized requires an active environment")
|
||||||
|
|
||||||
|
if env:
|
||||||
|
if args.constraint:
|
||||||
|
init_specs = spack.cmd.parse_specs(args.constraint)
|
||||||
|
results = env.all_matching_specs(*init_specs)
|
||||||
|
else:
|
||||||
|
results = env.all_specs()
|
||||||
|
else:
|
||||||
|
q_args = query_arguments(args)
|
||||||
|
results = args.specs(**q_args)
|
||||||
|
|
||||||
decorator = make_env_decorator(env) if env else lambda s, f: f
|
decorator = make_env_decorator(env) if env else lambda s, f: f
|
||||||
|
|
||||||
@@ -310,6 +335,11 @@ def find(parser, args):
|
|||||||
if args.loaded:
|
if args.loaded:
|
||||||
results = spack.cmd.filter_loaded_specs(results)
|
results = spack.cmd.filter_loaded_specs(results)
|
||||||
|
|
||||||
|
if args.install_status or args.show_concretized:
|
||||||
|
status_fn = spack.spec.Spec.install_status
|
||||||
|
else:
|
||||||
|
status_fn = None
|
||||||
|
|
||||||
# Display the result
|
# Display the result
|
||||||
if args.json:
|
if args.json:
|
||||||
cmd.display_specs_as_json(results, deps=args.deps)
|
cmd.display_specs_as_json(results, deps=args.deps)
|
||||||
@@ -318,12 +348,34 @@ def find(parser, args):
|
|||||||
if env:
|
if env:
|
||||||
display_env(env, args, decorator, results)
|
display_env(env, args, decorator, results)
|
||||||
|
|
||||||
count_suffix = " (not shown)"
|
|
||||||
if not args.only_roots:
|
if not args.only_roots:
|
||||||
cmd.display_specs(results, args, decorator=decorator, all_headers=True)
|
display_results = results
|
||||||
count_suffix = ""
|
if not args.show_concretized:
|
||||||
|
display_results = list(x for x in results if x.installed)
|
||||||
|
cmd.display_specs(
|
||||||
|
display_results, args, decorator=decorator, all_headers=True, status_fn=status_fn
|
||||||
|
)
|
||||||
|
|
||||||
# print number of installed packages last (as the list may be long)
|
# print number of installed packages last (as the list may be long)
|
||||||
if sys.stdout.isatty() and args.groups:
|
if sys.stdout.isatty() and args.groups:
|
||||||
|
installed_suffix = ""
|
||||||
|
concretized_suffix = " to be installed"
|
||||||
|
|
||||||
|
if args.only_roots:
|
||||||
|
installed_suffix += " (not shown)"
|
||||||
|
concretized_suffix += " (not shown)"
|
||||||
|
else:
|
||||||
|
if env and not args.show_concretized:
|
||||||
|
concretized_suffix += " (show with `spack find -c`)"
|
||||||
|
|
||||||
pkg_type = "loaded" if args.loaded else "installed"
|
pkg_type = "loaded" if args.loaded else "installed"
|
||||||
spack.cmd.print_how_many_pkgs(results, pkg_type, suffix=count_suffix)
|
spack.cmd.print_how_many_pkgs(
|
||||||
|
list(x for x in results if x.installed), pkg_type, suffix=installed_suffix
|
||||||
|
)
|
||||||
|
|
||||||
|
if env:
|
||||||
|
spack.cmd.print_how_many_pkgs(
|
||||||
|
list(x for x in results if not x.installed),
|
||||||
|
"concretized",
|
||||||
|
suffix=concretized_suffix,
|
||||||
|
)
|
||||||
|
|||||||
@@ -56,7 +56,6 @@ def roots_from_environments(args, active_env):
|
|||||||
|
|
||||||
# -e says "also preserve things needed by this particular env"
|
# -e says "also preserve things needed by this particular env"
|
||||||
for env_name_or_dir in args.except_environment:
|
for env_name_or_dir in args.except_environment:
|
||||||
print("HMM", env_name_or_dir)
|
|
||||||
if ev.exists(env_name_or_dir):
|
if ev.exists(env_name_or_dir):
|
||||||
env = ev.read(env_name_or_dir)
|
env = ev.read(env_name_or_dir)
|
||||||
elif ev.is_env_dir(env_name_or_dir):
|
elif ev.is_env_dir(env_name_or_dir):
|
||||||
|
|||||||
@@ -50,7 +50,7 @@
|
|||||||
@B{++}, @r{--}, @r{~~}, @B{==} propagate variants to package dependencies
|
@B{++}, @r{--}, @r{~~}, @B{==} propagate variants to package dependencies
|
||||||
|
|
||||||
architecture variants:
|
architecture variants:
|
||||||
@m{platform=platform} linux, darwin, cray, etc.
|
@m{platform=platform} linux, darwin, freebsd, windows
|
||||||
@m{os=operating_system} specific <operating_system>
|
@m{os=operating_system} specific <operating_system>
|
||||||
@m{target=target} specific <target> processor
|
@m{target=target} specific <target> processor
|
||||||
@m{arch=platform-os-target} shortcut for all three above
|
@m{arch=platform-os-target} shortcut for all three above
|
||||||
|
|||||||
@@ -10,6 +10,7 @@
|
|||||||
from typing import List
|
from typing import List
|
||||||
|
|
||||||
import llnl.util.filesystem as fs
|
import llnl.util.filesystem as fs
|
||||||
|
from llnl.string import plural
|
||||||
from llnl.util import lang, tty
|
from llnl.util import lang, tty
|
||||||
|
|
||||||
import spack.build_environment
|
import spack.build_environment
|
||||||
@@ -61,7 +62,6 @@ def install_kwargs_from_args(args):
|
|||||||
"dependencies_use_cache": cache_opt(args.use_cache, dep_use_bc),
|
"dependencies_use_cache": cache_opt(args.use_cache, dep_use_bc),
|
||||||
"dependencies_cache_only": cache_opt(args.cache_only, dep_use_bc),
|
"dependencies_cache_only": cache_opt(args.cache_only, dep_use_bc),
|
||||||
"include_build_deps": args.include_build_deps,
|
"include_build_deps": args.include_build_deps,
|
||||||
"explicit": True, # Use true as a default for install command
|
|
||||||
"stop_at": args.until,
|
"stop_at": args.until,
|
||||||
"unsigned": args.unsigned,
|
"unsigned": args.unsigned,
|
||||||
"install_deps": ("dependencies" in args.things_to_install),
|
"install_deps": ("dependencies" in args.things_to_install),
|
||||||
@@ -376,7 +376,9 @@ def _maybe_add_and_concretize(args, env, specs):
|
|||||||
# `spack concretize`
|
# `spack concretize`
|
||||||
tests = compute_tests_install_kwargs(env.user_specs, args.test)
|
tests = compute_tests_install_kwargs(env.user_specs, args.test)
|
||||||
concretized_specs = env.concretize(tests=tests)
|
concretized_specs = env.concretize(tests=tests)
|
||||||
ev.display_specs(concretized_specs)
|
if concretized_specs:
|
||||||
|
tty.msg(f"Concretized {plural(len(concretized_specs), 'spec')}")
|
||||||
|
ev.display_specs([concrete for _, concrete in concretized_specs])
|
||||||
|
|
||||||
# save view regeneration for later, so that we only do it
|
# save view regeneration for later, so that we only do it
|
||||||
# once, as it can be slow.
|
# once, as it can be slow.
|
||||||
@@ -473,6 +475,7 @@ def install_without_active_env(args, install_kwargs, reporter_factory):
|
|||||||
require_user_confirmation_for_overwrite(concrete_specs, args)
|
require_user_confirmation_for_overwrite(concrete_specs, args)
|
||||||
install_kwargs["overwrite"] = [spec.dag_hash() for spec in concrete_specs]
|
install_kwargs["overwrite"] = [spec.dag_hash() for spec in concrete_specs]
|
||||||
|
|
||||||
installs = [(s.package, install_kwargs) for s in concrete_specs]
|
installs = [s.package for s in concrete_specs]
|
||||||
builder = PackageInstaller(installs)
|
install_kwargs["explicit"] = [s.dag_hash() for s in concrete_specs]
|
||||||
|
builder = PackageInstaller(installs, install_kwargs)
|
||||||
builder.install()
|
builder.install()
|
||||||
|
|||||||
@@ -169,7 +169,9 @@ def pkg_hash(args):
|
|||||||
|
|
||||||
def get_grep(required=False):
|
def get_grep(required=False):
|
||||||
"""Get a grep command to use with ``spack pkg grep``."""
|
"""Get a grep command to use with ``spack pkg grep``."""
|
||||||
return exe.which(os.environ.get("SPACK_GREP") or "grep", required=required)
|
grep = exe.which(os.environ.get("SPACK_GREP") or "grep", required=required)
|
||||||
|
grep.ignore_quotes = True # allow `spack pkg grep '"quoted string"'` without warning
|
||||||
|
return grep
|
||||||
|
|
||||||
|
|
||||||
def pkg_grep(args, unknown_args):
|
def pkg_grep(args, unknown_args):
|
||||||
|
|||||||
@@ -91,7 +91,7 @@ def repo_add(args):
|
|||||||
tty.die("Not a Spack repository: %s" % path)
|
tty.die("Not a Spack repository: %s" % path)
|
||||||
|
|
||||||
# Make sure it's actually a spack repository by constructing it.
|
# Make sure it's actually a spack repository by constructing it.
|
||||||
repo = spack.repo.Repo(canon_path)
|
repo = spack.repo.from_path(canon_path)
|
||||||
|
|
||||||
# If that succeeds, finally add it to the configuration.
|
# If that succeeds, finally add it to the configuration.
|
||||||
repos = spack.config.get("repos", scope=args.scope)
|
repos = spack.config.get("repos", scope=args.scope)
|
||||||
@@ -124,7 +124,7 @@ def repo_remove(args):
|
|||||||
# If it is a namespace, remove corresponding repo
|
# If it is a namespace, remove corresponding repo
|
||||||
for path in repos:
|
for path in repos:
|
||||||
try:
|
try:
|
||||||
repo = spack.repo.Repo(path)
|
repo = spack.repo.from_path(path)
|
||||||
if repo.namespace == namespace_or_path:
|
if repo.namespace == namespace_or_path:
|
||||||
repos.remove(path)
|
repos.remove(path)
|
||||||
spack.config.set("repos", repos, args.scope)
|
spack.config.set("repos", repos, args.scope)
|
||||||
@@ -142,7 +142,7 @@ def repo_list(args):
|
|||||||
repos = []
|
repos = []
|
||||||
for r in roots:
|
for r in roots:
|
||||||
try:
|
try:
|
||||||
repos.append(spack.repo.Repo(r))
|
repos.append(spack.repo.from_path(r))
|
||||||
except spack.repo.RepoError:
|
except spack.repo.RepoError:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
|
|||||||
@@ -114,15 +114,16 @@ def _process_result(result, show, required_format, kwargs):
|
|||||||
|
|
||||||
# dump the solutions as concretized specs
|
# dump the solutions as concretized specs
|
||||||
if "solutions" in show:
|
if "solutions" in show:
|
||||||
for spec in result.specs:
|
if required_format:
|
||||||
# With -y, just print YAML to output.
|
for spec in result.specs:
|
||||||
if required_format == "yaml":
|
# With -y, just print YAML to output.
|
||||||
# use write because to_yaml already has a newline.
|
if required_format == "yaml":
|
||||||
sys.stdout.write(spec.to_yaml(hash=ht.dag_hash))
|
# use write because to_yaml already has a newline.
|
||||||
elif required_format == "json":
|
sys.stdout.write(spec.to_yaml(hash=ht.dag_hash))
|
||||||
sys.stdout.write(spec.to_json(hash=ht.dag_hash))
|
elif required_format == "json":
|
||||||
else:
|
sys.stdout.write(spec.to_json(hash=ht.dag_hash))
|
||||||
sys.stdout.write(spec.tree(color=sys.stdout.isatty(), **kwargs))
|
else:
|
||||||
|
sys.stdout.write(spack.spec.tree(result.specs, color=sys.stdout.isatty(), **kwargs))
|
||||||
print()
|
print()
|
||||||
|
|
||||||
if result.unsolved_specs and "solutions" in show:
|
if result.unsolved_specs and "solutions" in show:
|
||||||
|
|||||||
@@ -105,11 +105,19 @@ def spec(parser, args):
|
|||||||
if env:
|
if env:
|
||||||
env.concretize()
|
env.concretize()
|
||||||
specs = env.concretized_specs()
|
specs = env.concretized_specs()
|
||||||
|
|
||||||
|
# environments are printed together in a combined tree() invocation,
|
||||||
|
# except when using --yaml or --json, which we print spec by spec below.
|
||||||
|
if not args.format:
|
||||||
|
tree_kwargs["key"] = spack.traverse.by_dag_hash
|
||||||
|
tree_kwargs["hashes"] = args.long or args.very_long
|
||||||
|
print(spack.spec.tree([concrete for _, concrete in specs], **tree_kwargs))
|
||||||
|
return
|
||||||
else:
|
else:
|
||||||
tty.die("spack spec requires at least one spec or an active environment")
|
tty.die("spack spec requires at least one spec or an active environment")
|
||||||
|
|
||||||
for input, output in specs:
|
for input, output in specs:
|
||||||
# With -y, just print YAML to output.
|
# With --yaml or --json, just print the raw specs to output
|
||||||
if args.format:
|
if args.format:
|
||||||
if args.format == "yaml":
|
if args.format == "yaml":
|
||||||
# use write because to_yaml already has a newline.
|
# use write because to_yaml already has a newline.
|
||||||
|
|||||||
@@ -23,7 +23,7 @@
|
|||||||
|
|
||||||
|
|
||||||
# tutorial configuration parameters
|
# tutorial configuration parameters
|
||||||
tutorial_branch = "releases/v0.21"
|
tutorial_branch = "releases/v0.22"
|
||||||
tutorial_mirror = "file:///mirror"
|
tutorial_mirror = "file:///mirror"
|
||||||
tutorial_key = os.path.join(spack.paths.share_path, "keys", "tutorial.pub")
|
tutorial_key = os.path.join(spack.paths.share_path, "keys", "tutorial.pub")
|
||||||
|
|
||||||
|
|||||||
@@ -151,7 +151,8 @@ def is_installed(spec):
|
|||||||
key=lambda s: s.dag_hash(),
|
key=lambda s: s.dag_hash(),
|
||||||
)
|
)
|
||||||
|
|
||||||
return [spec for spec in specs if is_installed(spec)]
|
with spack.store.STORE.db.read_transaction():
|
||||||
|
return [spec for spec in specs if is_installed(spec)]
|
||||||
|
|
||||||
|
|
||||||
def dependent_environments(
|
def dependent_environments(
|
||||||
@@ -239,6 +240,8 @@ def get_uninstall_list(args, specs: List[spack.spec.Spec], env: Optional[ev.Envi
|
|||||||
print()
|
print()
|
||||||
tty.info("The following environments still reference these specs:")
|
tty.info("The following environments still reference these specs:")
|
||||||
colify([e.name for e in other_dependent_envs.keys()], indent=4)
|
colify([e.name for e in other_dependent_envs.keys()], indent=4)
|
||||||
|
if env:
|
||||||
|
msgs.append("use `spack remove` to remove the spec from the current environment")
|
||||||
msgs.append("use `spack env remove` to remove environments")
|
msgs.append("use `spack env remove` to remove environments")
|
||||||
msgs.append("use `spack uninstall --force` to override")
|
msgs.append("use `spack uninstall --force` to override")
|
||||||
print()
|
print()
|
||||||
|
|||||||
@@ -71,7 +71,7 @@ def unload(parser, args):
|
|||||||
"Cannot specify specs on command line when unloading all specs with '--all'"
|
"Cannot specify specs on command line when unloading all specs with '--all'"
|
||||||
)
|
)
|
||||||
|
|
||||||
hashes = os.environ.get(uenv.spack_loaded_hashes_var, "").split(":")
|
hashes = os.environ.get(uenv.spack_loaded_hashes_var, "").split(os.pathsep)
|
||||||
if args.specs:
|
if args.specs:
|
||||||
specs = [
|
specs = [
|
||||||
spack.cmd.disambiguate_spec_from_hashes(spec, hashes)
|
spack.cmd.disambiguate_spec_from_hashes(spec, hashes)
|
||||||
|
|||||||
@@ -38,10 +38,10 @@
|
|||||||
|
|
||||||
import spack.cmd
|
import spack.cmd
|
||||||
import spack.environment as ev
|
import spack.environment as ev
|
||||||
|
import spack.filesystem_view as fsv
|
||||||
import spack.schema.projections
|
import spack.schema.projections
|
||||||
import spack.store
|
import spack.store
|
||||||
from spack.config import validate
|
from spack.config import validate
|
||||||
from spack.filesystem_view import YamlFilesystemView, view_func_parser
|
|
||||||
from spack.util import spack_yaml as s_yaml
|
from spack.util import spack_yaml as s_yaml
|
||||||
|
|
||||||
description = "project packages to a compact naming scheme on the filesystem"
|
description = "project packages to a compact naming scheme on the filesystem"
|
||||||
@@ -193,17 +193,13 @@ def view(parser, args):
|
|||||||
ordered_projections = {}
|
ordered_projections = {}
|
||||||
|
|
||||||
# What method are we using for this view
|
# What method are we using for this view
|
||||||
if args.action in actions_link:
|
link_type = args.action if args.action in actions_link else "symlink"
|
||||||
link_fn = view_func_parser(args.action)
|
view = fsv.YamlFilesystemView(
|
||||||
else:
|
|
||||||
link_fn = view_func_parser("symlink")
|
|
||||||
|
|
||||||
view = YamlFilesystemView(
|
|
||||||
path,
|
path,
|
||||||
spack.store.STORE.layout,
|
spack.store.STORE.layout,
|
||||||
projections=ordered_projections,
|
projections=ordered_projections,
|
||||||
ignore_conflicts=getattr(args, "ignore_conflicts", False),
|
ignore_conflicts=getattr(args, "ignore_conflicts", False),
|
||||||
link=link_fn,
|
link_type=link_type,
|
||||||
verbose=args.verbose,
|
verbose=args.verbose,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -18,7 +18,6 @@
|
|||||||
import llnl.util.tty as tty
|
import llnl.util.tty as tty
|
||||||
from llnl.util.filesystem import path_contains_subdirectory, paths_containing_libs
|
from llnl.util.filesystem import path_contains_subdirectory, paths_containing_libs
|
||||||
|
|
||||||
import spack.compilers
|
|
||||||
import spack.error
|
import spack.error
|
||||||
import spack.schema.environment
|
import spack.schema.environment
|
||||||
import spack.spec
|
import spack.spec
|
||||||
@@ -695,10 +694,6 @@ def compiler_environment(self):
|
|||||||
try:
|
try:
|
||||||
# load modules and set env variables
|
# load modules and set env variables
|
||||||
for module in self.modules:
|
for module in self.modules:
|
||||||
# On cray, mic-knl module cannot be loaded without cce module
|
|
||||||
# See: https://github.com/spack/spack/issues/3153
|
|
||||||
if os.environ.get("CRAY_CPU_TARGET") == "mic-knl":
|
|
||||||
spack.util.module_cmd.load_module("cce")
|
|
||||||
spack.util.module_cmd.load_module(module)
|
spack.util.module_cmd.load_module(module)
|
||||||
|
|
||||||
# apply other compiler environment changes
|
# apply other compiler environment changes
|
||||||
|
|||||||
@@ -220,10 +220,10 @@ def _compiler_config_from_external(config):
|
|||||||
operating_system = host_platform.operating_system("default_os")
|
operating_system = host_platform.operating_system("default_os")
|
||||||
target = host_platform.target("default_target").microarchitecture
|
target = host_platform.target("default_target").microarchitecture
|
||||||
else:
|
else:
|
||||||
target = spec.target
|
target = spec.architecture.target
|
||||||
if not target:
|
if not target:
|
||||||
host_platform = spack.platforms.host()
|
target = spack.platforms.host().target("default_target")
|
||||||
target = host_platform.target("default_target").microarchitecture
|
target = target.microarchitecture
|
||||||
|
|
||||||
operating_system = spec.os
|
operating_system = spec.os
|
||||||
if not operating_system:
|
if not operating_system:
|
||||||
@@ -260,7 +260,7 @@ def _init_compiler_config(
|
|||||||
def compiler_config_files():
|
def compiler_config_files():
|
||||||
config_files = list()
|
config_files = list()
|
||||||
config = spack.config.CONFIG
|
config = spack.config.CONFIG
|
||||||
for scope in config.file_scopes:
|
for scope in config.writable_scopes:
|
||||||
name = scope.name
|
name = scope.name
|
||||||
compiler_config = config.get("compilers", scope=name)
|
compiler_config = config.get("compilers", scope=name)
|
||||||
if compiler_config:
|
if compiler_config:
|
||||||
@@ -488,7 +488,7 @@ def supported_compilers_for_host_platform() -> List[str]:
|
|||||||
return supported_compilers_for_platform(host_plat)
|
return supported_compilers_for_platform(host_plat)
|
||||||
|
|
||||||
|
|
||||||
def supported_compilers_for_platform(platform: spack.platforms.Platform) -> List[str]:
|
def supported_compilers_for_platform(platform: "spack.platforms.Platform") -> List[str]:
|
||||||
"""Return a set of compiler class objects supported by Spack
|
"""Return a set of compiler class objects supported by Spack
|
||||||
that are also supported by the provided platform
|
that are also supported by the provided platform
|
||||||
|
|
||||||
|
|||||||
@@ -96,6 +96,8 @@ def verbose_flag(self):
|
|||||||
|
|
||||||
openmp_flag = "-fopenmp"
|
openmp_flag = "-fopenmp"
|
||||||
|
|
||||||
|
# C++ flags based on CMake Modules/Compiler/Clang.cmake
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def cxx11_flag(self):
|
def cxx11_flag(self):
|
||||||
if self.real_version < Version("3.3"):
|
if self.real_version < Version("3.3"):
|
||||||
@@ -120,6 +122,24 @@ def cxx17_flag(self):
|
|||||||
|
|
||||||
return "-std=c++17"
|
return "-std=c++17"
|
||||||
|
|
||||||
|
@property
|
||||||
|
def cxx20_flag(self):
|
||||||
|
if self.real_version < Version("5.0"):
|
||||||
|
raise UnsupportedCompilerFlag(self, "the C++20 standard", "cxx20_flag", "< 5.0")
|
||||||
|
elif self.real_version < Version("11.0"):
|
||||||
|
return "-std=c++2a"
|
||||||
|
else:
|
||||||
|
return "-std=c++20"
|
||||||
|
|
||||||
|
@property
|
||||||
|
def cxx23_flag(self):
|
||||||
|
if self.real_version < Version("12.0"):
|
||||||
|
raise UnsupportedCompilerFlag(self, "the C++23 standard", "cxx23_flag", "< 12.0")
|
||||||
|
elif self.real_version < Version("17.0"):
|
||||||
|
return "-std=c++2b"
|
||||||
|
else:
|
||||||
|
return "-std=c++23"
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def c99_flag(self):
|
def c99_flag(self):
|
||||||
return "-std=c99"
|
return "-std=c99"
|
||||||
@@ -142,7 +162,10 @@ def c17_flag(self):
|
|||||||
def c23_flag(self):
|
def c23_flag(self):
|
||||||
if self.real_version < Version("9.0"):
|
if self.real_version < Version("9.0"):
|
||||||
raise UnsupportedCompilerFlag(self, "the C23 standard", "c23_flag", "< 9.0")
|
raise UnsupportedCompilerFlag(self, "the C23 standard", "c23_flag", "< 9.0")
|
||||||
return "-std=c2x"
|
elif self.real_version < Version("18.0"):
|
||||||
|
return "-std=c2x"
|
||||||
|
else:
|
||||||
|
return "-std=c23"
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def cc_pic_flag(self):
|
def cc_pic_flag(self):
|
||||||
|
|||||||
@@ -4,7 +4,7 @@
|
|||||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
||||||
|
|
||||||
import os
|
import os
|
||||||
from os.path import dirname
|
from os.path import dirname, join
|
||||||
|
|
||||||
from llnl.util import tty
|
from llnl.util import tty
|
||||||
|
|
||||||
@@ -135,8 +135,12 @@ def setup_custom_environment(self, pkg, env):
|
|||||||
# It is located in the same directory as the driver. Error message:
|
# It is located in the same directory as the driver. Error message:
|
||||||
# clang++: error: unable to execute command:
|
# clang++: error: unable to execute command:
|
||||||
# Executable "sycl-post-link" doesn't exist!
|
# Executable "sycl-post-link" doesn't exist!
|
||||||
if self.cxx:
|
# also ensures that shared objects and libraries required by the compiler,
|
||||||
|
# e.g. libonnx, can be found succesfully
|
||||||
|
# due to a fix, this is no longer required for OneAPI versions >= 2024.2
|
||||||
|
if self.cxx and pkg.spec.satisfies("%oneapi@:2024.1"):
|
||||||
env.prepend_path("PATH", dirname(self.cxx))
|
env.prepend_path("PATH", dirname(self.cxx))
|
||||||
|
env.prepend_path("LD_LIBRARY_PATH", join(dirname(dirname(self.cxx)), "lib"))
|
||||||
|
|
||||||
# 2024 release bumped the libsycl version because of an ABI
|
# 2024 release bumped the libsycl version because of an ABI
|
||||||
# change, 2024 compilers are required. You will see this
|
# change, 2024 compilers are required. You will see this
|
||||||
|
|||||||
@@ -35,11 +35,10 @@
|
|||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import sys
|
import sys
|
||||||
from typing import Any, Callable, Dict, Generator, List, Optional, Tuple, Type, Union
|
from typing import Any, Callable, Dict, Generator, List, Optional, Tuple, Union
|
||||||
|
|
||||||
from llnl.util import filesystem, lang, tty
|
from llnl.util import filesystem, lang, tty
|
||||||
|
|
||||||
import spack.compilers
|
|
||||||
import spack.paths
|
import spack.paths
|
||||||
import spack.platforms
|
import spack.platforms
|
||||||
import spack.schema
|
import spack.schema
|
||||||
@@ -117,21 +116,39 @@
|
|||||||
|
|
||||||
|
|
||||||
class ConfigScope:
|
class ConfigScope:
|
||||||
"""This class represents a configuration scope.
|
def __init__(self, name: str) -> None:
|
||||||
|
self.name = name
|
||||||
|
self.writable = False
|
||||||
|
self.sections = syaml.syaml_dict()
|
||||||
|
|
||||||
A scope is one directory containing named configuration files.
|
def get_section_filename(self, section: str) -> str:
|
||||||
Each file is a config "section" (e.g., mirrors, compilers, etc.).
|
raise NotImplementedError
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, name, path) -> None:
|
def get_section(self, section: str) -> Optional[YamlConfigDict]:
|
||||||
self.name = name # scope name.
|
raise NotImplementedError
|
||||||
self.path = path # path to directory containing configs.
|
|
||||||
self.sections = syaml.syaml_dict() # sections read from config files.
|
def _write_section(self, section: str) -> None:
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def is_platform_dependent(self) -> bool:
|
def is_platform_dependent(self) -> bool:
|
||||||
"""Returns true if the scope name is platform specific"""
|
return False
|
||||||
return os.sep in self.name
|
|
||||||
|
def clear(self) -> None:
|
||||||
|
"""Empty cached config information."""
|
||||||
|
self.sections = syaml.syaml_dict()
|
||||||
|
|
||||||
|
def __repr__(self) -> str:
|
||||||
|
return f"<ConfigScope: {self.name}>"
|
||||||
|
|
||||||
|
|
||||||
|
class DirectoryConfigScope(ConfigScope):
|
||||||
|
"""Config scope backed by a directory containing one file per section."""
|
||||||
|
|
||||||
|
def __init__(self, name: str, path: str, *, writable: bool = True) -> None:
|
||||||
|
super().__init__(name)
|
||||||
|
self.path = path
|
||||||
|
self.writable = writable
|
||||||
|
|
||||||
def get_section_filename(self, section: str) -> str:
|
def get_section_filename(self, section: str) -> str:
|
||||||
"""Returns the filename associated with a given section"""
|
"""Returns the filename associated with a given section"""
|
||||||
@@ -148,14 +165,15 @@ def get_section(self, section: str) -> Optional[YamlConfigDict]:
|
|||||||
return self.sections[section]
|
return self.sections[section]
|
||||||
|
|
||||||
def _write_section(self, section: str) -> None:
|
def _write_section(self, section: str) -> None:
|
||||||
|
if not self.writable:
|
||||||
|
raise ConfigError(f"Cannot write to immutable scope {self}")
|
||||||
|
|
||||||
filename = self.get_section_filename(section)
|
filename = self.get_section_filename(section)
|
||||||
data = self.get_section(section)
|
data = self.get_section(section)
|
||||||
if data is None:
|
if data is None:
|
||||||
return
|
return
|
||||||
|
|
||||||
# We copy data here to avoid adding defaults at write time
|
validate(data, SECTION_SCHEMAS[section])
|
||||||
validate_data = copy.deepcopy(data)
|
|
||||||
validate(validate_data, SECTION_SCHEMAS[section])
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
filesystem.mkdirp(self.path)
|
filesystem.mkdirp(self.path)
|
||||||
@@ -164,19 +182,23 @@ def _write_section(self, section: str) -> None:
|
|||||||
except (syaml.SpackYAMLError, OSError) as e:
|
except (syaml.SpackYAMLError, OSError) as e:
|
||||||
raise ConfigFileError(f"cannot write to '{filename}'") from e
|
raise ConfigFileError(f"cannot write to '{filename}'") from e
|
||||||
|
|
||||||
def clear(self) -> None:
|
@property
|
||||||
"""Empty cached config information."""
|
def is_platform_dependent(self) -> bool:
|
||||||
self.sections = syaml.syaml_dict()
|
"""Returns true if the scope name is platform specific"""
|
||||||
|
return "/" in self.name
|
||||||
def __repr__(self) -> str:
|
|
||||||
return f"<ConfigScope: {self.name}: {self.path}>"
|
|
||||||
|
|
||||||
|
|
||||||
class SingleFileScope(ConfigScope):
|
class SingleFileScope(ConfigScope):
|
||||||
"""This class represents a configuration scope in a single YAML file."""
|
"""This class represents a configuration scope in a single YAML file."""
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self, name: str, path: str, schema: YamlConfigDict, yaml_path: Optional[List[str]] = None
|
self,
|
||||||
|
name: str,
|
||||||
|
path: str,
|
||||||
|
schema: YamlConfigDict,
|
||||||
|
*,
|
||||||
|
yaml_path: Optional[List[str]] = None,
|
||||||
|
writable: bool = True,
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Similar to ``ConfigScope`` but can be embedded in another schema.
|
"""Similar to ``ConfigScope`` but can be embedded in another schema.
|
||||||
|
|
||||||
@@ -195,15 +217,13 @@ def __init__(
|
|||||||
config:
|
config:
|
||||||
install_tree: $spack/opt/spack
|
install_tree: $spack/opt/spack
|
||||||
"""
|
"""
|
||||||
super().__init__(name, path)
|
super().__init__(name)
|
||||||
self._raw_data: Optional[YamlConfigDict] = None
|
self._raw_data: Optional[YamlConfigDict] = None
|
||||||
self.schema = schema
|
self.schema = schema
|
||||||
|
self.path = path
|
||||||
|
self.writable = writable
|
||||||
self.yaml_path = yaml_path or []
|
self.yaml_path = yaml_path or []
|
||||||
|
|
||||||
@property
|
|
||||||
def is_platform_dependent(self) -> bool:
|
|
||||||
return False
|
|
||||||
|
|
||||||
def get_section_filename(self, section) -> str:
|
def get_section_filename(self, section) -> str:
|
||||||
return self.path
|
return self.path
|
||||||
|
|
||||||
@@ -257,6 +277,8 @@ def get_section(self, section: str) -> Optional[YamlConfigDict]:
|
|||||||
return self.sections.get(section, None)
|
return self.sections.get(section, None)
|
||||||
|
|
||||||
def _write_section(self, section: str) -> None:
|
def _write_section(self, section: str) -> None:
|
||||||
|
if not self.writable:
|
||||||
|
raise ConfigError(f"Cannot write to immutable scope {self}")
|
||||||
data_to_write: Optional[YamlConfigDict] = self._raw_data
|
data_to_write: Optional[YamlConfigDict] = self._raw_data
|
||||||
|
|
||||||
# If there is no existing data, this section SingleFileScope has never
|
# If there is no existing data, this section SingleFileScope has never
|
||||||
@@ -301,19 +323,6 @@ def __repr__(self) -> str:
|
|||||||
return f"<SingleFileScope: {self.name}: {self.path}>"
|
return f"<SingleFileScope: {self.name}: {self.path}>"
|
||||||
|
|
||||||
|
|
||||||
class ImmutableConfigScope(ConfigScope):
|
|
||||||
"""A configuration scope that cannot be written to.
|
|
||||||
|
|
||||||
This is used for ConfigScopes passed on the command line.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def _write_section(self, section) -> None:
|
|
||||||
raise ConfigError(f"Cannot write to immutable scope {self}")
|
|
||||||
|
|
||||||
def __repr__(self) -> str:
|
|
||||||
return f"<ImmutableConfigScope: {self.name}: {self.path}>"
|
|
||||||
|
|
||||||
|
|
||||||
class InternalConfigScope(ConfigScope):
|
class InternalConfigScope(ConfigScope):
|
||||||
"""An internal configuration scope that is not persisted to a file.
|
"""An internal configuration scope that is not persisted to a file.
|
||||||
|
|
||||||
@@ -323,7 +332,7 @@ class InternalConfigScope(ConfigScope):
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, name: str, data: Optional[YamlConfigDict] = None) -> None:
|
def __init__(self, name: str, data: Optional[YamlConfigDict] = None) -> None:
|
||||||
super().__init__(name, None)
|
super().__init__(name)
|
||||||
self.sections = syaml.syaml_dict()
|
self.sections = syaml.syaml_dict()
|
||||||
|
|
||||||
if data is not None:
|
if data is not None:
|
||||||
@@ -333,9 +342,6 @@ def __init__(self, name: str, data: Optional[YamlConfigDict] = None) -> None:
|
|||||||
validate({section: dsec}, SECTION_SCHEMAS[section])
|
validate({section: dsec}, SECTION_SCHEMAS[section])
|
||||||
self.sections[section] = _mark_internal(syaml.syaml_dict({section: dsec}), name)
|
self.sections[section] = _mark_internal(syaml.syaml_dict({section: dsec}), name)
|
||||||
|
|
||||||
def get_section_filename(self, section: str) -> str:
|
|
||||||
raise NotImplementedError("Cannot get filename for InternalConfigScope.")
|
|
||||||
|
|
||||||
def get_section(self, section: str) -> Optional[YamlConfigDict]:
|
def get_section(self, section: str) -> Optional[YamlConfigDict]:
|
||||||
"""Just reads from an internal dictionary."""
|
"""Just reads from an internal dictionary."""
|
||||||
if section not in self.sections:
|
if section not in self.sections:
|
||||||
@@ -440,27 +446,21 @@ def remove_scope(self, scope_name: str) -> Optional[ConfigScope]:
|
|||||||
return scope
|
return scope
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def file_scopes(self) -> List[ConfigScope]:
|
def writable_scopes(self) -> Generator[ConfigScope, None, None]:
|
||||||
"""List of writable scopes with an associated file."""
|
"""Generator of writable scopes with an associated file."""
|
||||||
return [
|
return (s for s in self.scopes.values() if s.writable)
|
||||||
s
|
|
||||||
for s in self.scopes.values()
|
|
||||||
if (type(s) is ConfigScope or type(s) is SingleFileScope)
|
|
||||||
]
|
|
||||||
|
|
||||||
def highest_precedence_scope(self) -> ConfigScope:
|
def highest_precedence_scope(self) -> ConfigScope:
|
||||||
"""Non-internal scope with highest precedence."""
|
"""Writable scope with highest precedence."""
|
||||||
return next(reversed(self.file_scopes))
|
return next(s for s in reversed(self.scopes.values()) if s.writable) # type: ignore
|
||||||
|
|
||||||
def highest_precedence_non_platform_scope(self) -> ConfigScope:
|
def highest_precedence_non_platform_scope(self) -> ConfigScope:
|
||||||
"""Non-internal non-platform scope with highest precedence
|
"""Writable non-platform scope with highest precedence"""
|
||||||
|
return next(
|
||||||
Platform-specific scopes are of the form scope/platform"""
|
s
|
||||||
generator = reversed(self.file_scopes)
|
for s in reversed(self.scopes.values()) # type: ignore
|
||||||
highest = next(generator)
|
if s.writable and not s.is_platform_dependent
|
||||||
while highest and highest.is_platform_dependent:
|
)
|
||||||
highest = next(generator)
|
|
||||||
return highest
|
|
||||||
|
|
||||||
def matching_scopes(self, reg_expr) -> List[ConfigScope]:
|
def matching_scopes(self, reg_expr) -> List[ConfigScope]:
|
||||||
"""
|
"""
|
||||||
@@ -755,13 +755,14 @@ def override(
|
|||||||
|
|
||||||
|
|
||||||
def _add_platform_scope(
|
def _add_platform_scope(
|
||||||
cfg: Union[Configuration, lang.Singleton], scope_type: Type[ConfigScope], name: str, path: str
|
cfg: Union[Configuration, lang.Singleton], name: str, path: str, writable: bool = True
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Add a platform-specific subdirectory for the current platform."""
|
"""Add a platform-specific subdirectory for the current platform."""
|
||||||
platform = spack.platforms.host().name
|
platform = spack.platforms.host().name
|
||||||
plat_name = os.path.join(name, platform)
|
scope = DirectoryConfigScope(
|
||||||
plat_path = os.path.join(path, platform)
|
f"{name}/{platform}", os.path.join(path, platform), writable=writable
|
||||||
cfg.push_scope(scope_type(plat_name, plat_path))
|
)
|
||||||
|
cfg.push_scope(scope)
|
||||||
|
|
||||||
|
|
||||||
def config_paths_from_entry_points() -> List[Tuple[str, str]]:
|
def config_paths_from_entry_points() -> List[Tuple[str, str]]:
|
||||||
@@ -792,22 +793,27 @@ def config_paths_from_entry_points() -> List[Tuple[str, str]]:
|
|||||||
def _add_command_line_scopes(
|
def _add_command_line_scopes(
|
||||||
cfg: Union[Configuration, lang.Singleton], command_line_scopes: List[str]
|
cfg: Union[Configuration, lang.Singleton], command_line_scopes: List[str]
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Add additional scopes from the --config-scope argument.
|
"""Add additional scopes from the --config-scope argument, either envs or dirs."""
|
||||||
|
import spack.environment.environment as env # circular import
|
||||||
|
|
||||||
Command line scopes are named after their position in the arg list.
|
|
||||||
"""
|
|
||||||
for i, path in enumerate(command_line_scopes):
|
for i, path in enumerate(command_line_scopes):
|
||||||
# We ensure that these scopes exist and are readable, as they are
|
name = f"cmd_scope_{i}"
|
||||||
# provided on the command line by the user.
|
|
||||||
if not os.path.isdir(path):
|
|
||||||
raise ConfigError(f"config scope is not a directory: '{path}'")
|
|
||||||
elif not os.access(path, os.R_OK):
|
|
||||||
raise ConfigError(f"config scope is not readable: '{path}'")
|
|
||||||
|
|
||||||
# name based on order on the command line
|
if env.exists(path): # managed environment
|
||||||
name = f"cmd_scope_{i:d}"
|
manifest = env.EnvironmentManifestFile(env.root(path))
|
||||||
cfg.push_scope(ImmutableConfigScope(name, path))
|
elif env.is_env_dir(path): # anonymous environment
|
||||||
_add_platform_scope(cfg, ImmutableConfigScope, name, path)
|
manifest = env.EnvironmentManifestFile(path)
|
||||||
|
elif os.path.isdir(path): # directory with config files
|
||||||
|
cfg.push_scope(DirectoryConfigScope(name, path, writable=False))
|
||||||
|
_add_platform_scope(cfg, name, path, writable=False)
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
raise ConfigError(f"Invalid configuration scope: {path}")
|
||||||
|
|
||||||
|
for scope in manifest.env_config_scopes:
|
||||||
|
scope.name = f"{name}:{scope.name}"
|
||||||
|
scope.writable = False
|
||||||
|
cfg.push_scope(scope)
|
||||||
|
|
||||||
|
|
||||||
def create() -> Configuration:
|
def create() -> Configuration:
|
||||||
@@ -851,10 +857,10 @@ def create() -> Configuration:
|
|||||||
|
|
||||||
# add each scope and its platform-specific directory
|
# add each scope and its platform-specific directory
|
||||||
for name, path in configuration_paths:
|
for name, path in configuration_paths:
|
||||||
cfg.push_scope(ConfigScope(name, path))
|
cfg.push_scope(DirectoryConfigScope(name, path))
|
||||||
|
|
||||||
# Each scope can have per-platfom overrides in subdirectories
|
# Each scope can have per-platfom overrides in subdirectories
|
||||||
_add_platform_scope(cfg, ConfigScope, name, path)
|
_add_platform_scope(cfg, name, path)
|
||||||
|
|
||||||
# add command-line scopes
|
# add command-line scopes
|
||||||
_add_command_line_scopes(cfg, COMMAND_LINE_SCOPES)
|
_add_command_line_scopes(cfg, COMMAND_LINE_SCOPES)
|
||||||
@@ -969,7 +975,7 @@ def set(path: str, value: Any, scope: Optional[str] = None) -> None:
|
|||||||
def add_default_platform_scope(platform: str) -> None:
|
def add_default_platform_scope(platform: str) -> None:
|
||||||
plat_name = os.path.join("defaults", platform)
|
plat_name = os.path.join("defaults", platform)
|
||||||
plat_path = os.path.join(CONFIGURATION_DEFAULTS_PATH[1], platform)
|
plat_path = os.path.join(CONFIGURATION_DEFAULTS_PATH[1], platform)
|
||||||
CONFIG.push_scope(ConfigScope(plat_name, plat_path))
|
CONFIG.push_scope(DirectoryConfigScope(plat_name, plat_path))
|
||||||
|
|
||||||
|
|
||||||
def scopes() -> Dict[str, ConfigScope]:
|
def scopes() -> Dict[str, ConfigScope]:
|
||||||
@@ -978,19 +984,10 @@ def scopes() -> Dict[str, ConfigScope]:
|
|||||||
|
|
||||||
|
|
||||||
def writable_scopes() -> List[ConfigScope]:
|
def writable_scopes() -> List[ConfigScope]:
|
||||||
"""
|
"""Return list of writable scopes. Higher-priority scopes come first in the list."""
|
||||||
Return list of writable scopes. Higher-priority scopes come first in the
|
scopes = [x for x in CONFIG.scopes.values() if x.writable]
|
||||||
list.
|
scopes.reverse()
|
||||||
"""
|
return scopes
|
||||||
return list(
|
|
||||||
reversed(
|
|
||||||
list(
|
|
||||||
x
|
|
||||||
for x in CONFIG.scopes.values()
|
|
||||||
if not isinstance(x, (InternalConfigScope, ImmutableConfigScope))
|
|
||||||
)
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def writable_scope_names() -> List[str]:
|
def writable_scope_names() -> List[str]:
|
||||||
@@ -1080,11 +1077,8 @@ def validate(
|
|||||||
"""
|
"""
|
||||||
import jsonschema
|
import jsonschema
|
||||||
|
|
||||||
# Validate a copy to avoid adding defaults
|
|
||||||
# This allows us to round-trip data without adding to it.
|
|
||||||
test_data = syaml.deepcopy(data)
|
|
||||||
try:
|
try:
|
||||||
spack.schema.Validator(schema).validate(test_data)
|
spack.schema.Validator(schema).validate(data)
|
||||||
except jsonschema.ValidationError as e:
|
except jsonschema.ValidationError as e:
|
||||||
if hasattr(e.instance, "lc"):
|
if hasattr(e.instance, "lc"):
|
||||||
line_number = e.instance.lc.line + 1
|
line_number = e.instance.lc.line + 1
|
||||||
@@ -1093,7 +1087,7 @@ def validate(
|
|||||||
raise ConfigFormatError(e, data, filename, line_number) from e
|
raise ConfigFormatError(e, data, filename, line_number) from e
|
||||||
# return the validated data so that we can access the raw data
|
# return the validated data so that we can access the raw data
|
||||||
# mostly relevant for environments
|
# mostly relevant for environments
|
||||||
return test_data
|
return data
|
||||||
|
|
||||||
|
|
||||||
def read_config_file(
|
def read_config_file(
|
||||||
@@ -1599,7 +1593,7 @@ def _config_from(scopes_or_paths: List[Union[ConfigScope, str]]) -> Configuratio
|
|||||||
path = os.path.normpath(scope_or_path)
|
path = os.path.normpath(scope_or_path)
|
||||||
assert os.path.isdir(path), f'"{path}" must be a directory'
|
assert os.path.isdir(path), f'"{path}" must be a directory'
|
||||||
name = os.path.basename(path)
|
name = os.path.basename(path)
|
||||||
scopes.append(ConfigScope(name, path))
|
scopes.append(DirectoryConfigScope(name, path))
|
||||||
|
|
||||||
configuration = Configuration(*scopes)
|
configuration = Configuration(*scopes)
|
||||||
return configuration
|
return configuration
|
||||||
|
|||||||
@@ -78,24 +78,17 @@
|
|||||||
"image": "quay.io/almalinuxorg/almalinux:8"
|
"image": "quay.io/almalinuxorg/almalinux:8"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"centos:stream": {
|
"centos:stream9": {
|
||||||
"bootstrap": {
|
"bootstrap": {
|
||||||
"template": "container/centos_stream.dockerfile",
|
"template": "container/centos_stream9.dockerfile",
|
||||||
"image": "quay.io/centos/centos:stream"
|
"image": "quay.io/centos/centos:stream9"
|
||||||
},
|
},
|
||||||
"os_package_manager": "dnf_epel",
|
"os_package_manager": "dnf_epel",
|
||||||
"build": "spack/centos-stream",
|
"build": "spack/centos-stream9",
|
||||||
"final": {
|
"final": {
|
||||||
"image": "quay.io/centos/centos:stream"
|
"image": "quay.io/centos/centos:stream9"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"centos:7": {
|
|
||||||
"bootstrap": {
|
|
||||||
"template": "container/centos_7.dockerfile"
|
|
||||||
},
|
|
||||||
"os_package_manager": "yum",
|
|
||||||
"build": "spack/centos7"
|
|
||||||
},
|
|
||||||
"opensuse/leap:15": {
|
"opensuse/leap:15": {
|
||||||
"bootstrap": {
|
"bootstrap": {
|
||||||
"template": "container/leap-15.dockerfile"
|
"template": "container/leap-15.dockerfile"
|
||||||
|
|||||||
@@ -2,7 +2,12 @@
|
|||||||
# Spack Project Developers. See the top-level COPYRIGHT file for details.
|
# Spack Project Developers. See the top-level COPYRIGHT file for details.
|
||||||
#
|
#
|
||||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
||||||
from .common import DetectedPackage, executable_prefix, update_configuration
|
from .common import (
|
||||||
|
DetectedPackage,
|
||||||
|
executable_prefix,
|
||||||
|
set_virtuals_nonbuildable,
|
||||||
|
update_configuration,
|
||||||
|
)
|
||||||
from .path import by_path, executables_in_path
|
from .path import by_path, executables_in_path
|
||||||
from .test import detection_tests
|
from .test import detection_tests
|
||||||
|
|
||||||
@@ -12,5 +17,6 @@
|
|||||||
"executables_in_path",
|
"executables_in_path",
|
||||||
"executable_prefix",
|
"executable_prefix",
|
||||||
"update_configuration",
|
"update_configuration",
|
||||||
|
"set_virtuals_nonbuildable",
|
||||||
"detection_tests",
|
"detection_tests",
|
||||||
]
|
]
|
||||||
|
|||||||
@@ -252,6 +252,27 @@ def update_configuration(
|
|||||||
return all_new_specs
|
return all_new_specs
|
||||||
|
|
||||||
|
|
||||||
|
def set_virtuals_nonbuildable(virtuals: Set[str], scope: Optional[str] = None) -> List[str]:
|
||||||
|
"""Update packages:virtual:buildable:False for the provided virtual packages, if the property
|
||||||
|
is not set by the user. Returns the list of virtual packages that have been updated."""
|
||||||
|
packages = spack.config.get("packages")
|
||||||
|
new_config = {}
|
||||||
|
for virtual in virtuals:
|
||||||
|
# If the user has set the buildable prop do not override it
|
||||||
|
if virtual in packages and "buildable" in packages[virtual]:
|
||||||
|
continue
|
||||||
|
new_config[virtual] = {"buildable": False}
|
||||||
|
|
||||||
|
# Update the provided scope
|
||||||
|
spack.config.set(
|
||||||
|
"packages",
|
||||||
|
spack.config.merge_yaml(spack.config.get("packages", scope=scope), new_config),
|
||||||
|
scope=scope,
|
||||||
|
)
|
||||||
|
|
||||||
|
return list(new_config.keys())
|
||||||
|
|
||||||
|
|
||||||
def _windows_drive() -> str:
|
def _windows_drive() -> str:
|
||||||
"""Return Windows drive string extracted from the PROGRAMFILES environment variable,
|
"""Return Windows drive string extracted from the PROGRAMFILES environment variable,
|
||||||
which is guaranteed to be defined for all logins.
|
which is guaranteed to be defined for all logins.
|
||||||
|
|||||||
@@ -12,7 +12,7 @@
|
|||||||
import re
|
import re
|
||||||
import sys
|
import sys
|
||||||
import warnings
|
import warnings
|
||||||
from typing import Dict, List, Optional, Set, Tuple
|
from typing import Dict, List, Optional, Set, Tuple, Type
|
||||||
|
|
||||||
import llnl.util.filesystem
|
import llnl.util.filesystem
|
||||||
import llnl.util.lang
|
import llnl.util.lang
|
||||||
@@ -200,7 +200,7 @@ class Finder:
|
|||||||
def default_path_hints(self) -> List[str]:
|
def default_path_hints(self) -> List[str]:
|
||||||
return []
|
return []
|
||||||
|
|
||||||
def search_patterns(self, *, pkg: "spack.package_base.PackageBase") -> List[str]:
|
def search_patterns(self, *, pkg: Type["spack.package_base.PackageBase"]) -> List[str]:
|
||||||
"""Returns the list of patterns used to match candidate files.
|
"""Returns the list of patterns used to match candidate files.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
@@ -226,7 +226,7 @@ def prefix_from_path(self, *, path: str) -> str:
|
|||||||
raise NotImplementedError("must be implemented by derived classes")
|
raise NotImplementedError("must be implemented by derived classes")
|
||||||
|
|
||||||
def detect_specs(
|
def detect_specs(
|
||||||
self, *, pkg: "spack.package_base.PackageBase", paths: List[str]
|
self, *, pkg: Type["spack.package_base.PackageBase"], paths: List[str]
|
||||||
) -> List[DetectedPackage]:
|
) -> List[DetectedPackage]:
|
||||||
"""Given a list of files matching the search patterns, returns a list of detected specs.
|
"""Given a list of files matching the search patterns, returns a list of detected specs.
|
||||||
|
|
||||||
@@ -327,7 +327,7 @@ class ExecutablesFinder(Finder):
|
|||||||
def default_path_hints(self) -> List[str]:
|
def default_path_hints(self) -> List[str]:
|
||||||
return spack.util.environment.get_path("PATH")
|
return spack.util.environment.get_path("PATH")
|
||||||
|
|
||||||
def search_patterns(self, *, pkg: "spack.package_base.PackageBase") -> List[str]:
|
def search_patterns(self, *, pkg: Type["spack.package_base.PackageBase"]) -> List[str]:
|
||||||
result = []
|
result = []
|
||||||
if hasattr(pkg, "executables") and hasattr(pkg, "platform_executables"):
|
if hasattr(pkg, "executables") and hasattr(pkg, "platform_executables"):
|
||||||
result = pkg.platform_executables()
|
result = pkg.platform_executables()
|
||||||
@@ -356,7 +356,7 @@ class LibrariesFinder(Finder):
|
|||||||
DYLD_LIBRARY_PATH, DYLD_FALLBACK_LIBRARY_PATH, and standard system library paths
|
DYLD_LIBRARY_PATH, DYLD_FALLBACK_LIBRARY_PATH, and standard system library paths
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def search_patterns(self, *, pkg: "spack.package_base.PackageBase") -> List[str]:
|
def search_patterns(self, *, pkg: Type["spack.package_base.PackageBase"]) -> List[str]:
|
||||||
result = []
|
result = []
|
||||||
if hasattr(pkg, "libraries"):
|
if hasattr(pkg, "libraries"):
|
||||||
result = pkg.libraries
|
result = pkg.libraries
|
||||||
|
|||||||
@@ -90,14 +90,14 @@ class OpenMpi(Package):
|
|||||||
_patch_order_index = 0
|
_patch_order_index = 0
|
||||||
|
|
||||||
|
|
||||||
SpecType = Union["spack.spec.Spec", str]
|
SpecType = str
|
||||||
DepType = Union[Tuple[str, ...], str]
|
DepType = Union[Tuple[str, ...], str]
|
||||||
WhenType = Optional[Union["spack.spec.Spec", str, bool]]
|
WhenType = Optional[Union["spack.spec.Spec", str, bool]]
|
||||||
Patcher = Callable[[Union["spack.package_base.PackageBase", Dependency]], None]
|
Patcher = Callable[[Union["spack.package_base.PackageBase", Dependency]], None]
|
||||||
PatchesType = Optional[Union[Patcher, str, List[Union[Patcher, str]]]]
|
PatchesType = Optional[Union[Patcher, str, List[Union[Patcher, str]]]]
|
||||||
|
|
||||||
|
|
||||||
SUPPORTED_LANGUAGES = ("fortran", "cxx")
|
SUPPORTED_LANGUAGES = ("fortran", "cxx", "c")
|
||||||
|
|
||||||
|
|
||||||
def _make_when_spec(value: WhenType) -> Optional["spack.spec.Spec"]:
|
def _make_when_spec(value: WhenType) -> Optional["spack.spec.Spec"]:
|
||||||
@@ -475,7 +475,7 @@ def _execute_version(pkg, ver, **kwargs):
|
|||||||
|
|
||||||
def _depends_on(
|
def _depends_on(
|
||||||
pkg: "spack.package_base.PackageBase",
|
pkg: "spack.package_base.PackageBase",
|
||||||
spec: SpecType,
|
spec: "spack.spec.Spec",
|
||||||
*,
|
*,
|
||||||
when: WhenType = None,
|
when: WhenType = None,
|
||||||
type: DepType = dt.DEFAULT_TYPES,
|
type: DepType = dt.DEFAULT_TYPES,
|
||||||
@@ -485,11 +485,10 @@ def _depends_on(
|
|||||||
if not when_spec:
|
if not when_spec:
|
||||||
return
|
return
|
||||||
|
|
||||||
dep_spec = spack.spec.Spec(spec)
|
if not spec.name:
|
||||||
if not dep_spec.name:
|
raise DependencyError(f"Invalid dependency specification in package '{pkg.name}':", spec)
|
||||||
raise DependencyError("Invalid dependency specification in package '%s':" % pkg.name, spec)
|
if pkg.name == spec.name:
|
||||||
if pkg.name == dep_spec.name:
|
raise CircularReferenceError(f"Package '{pkg.name}' cannot depend on itself.")
|
||||||
raise CircularReferenceError("Package '%s' cannot depend on itself." % pkg.name)
|
|
||||||
|
|
||||||
depflag = dt.canonicalize(type)
|
depflag = dt.canonicalize(type)
|
||||||
|
|
||||||
@@ -505,7 +504,7 @@ def _depends_on(
|
|||||||
# ensure `Spec.virtual` is a valid thing to call in a directive.
|
# ensure `Spec.virtual` is a valid thing to call in a directive.
|
||||||
# For now, we comment out the following check to allow for virtual packages
|
# For now, we comment out the following check to allow for virtual packages
|
||||||
# with package files.
|
# with package files.
|
||||||
# if patches and dep_spec.virtual:
|
# if patches and spec.virtual:
|
||||||
# raise DependencyPatchError("Cannot patch a virtual dependency.")
|
# raise DependencyPatchError("Cannot patch a virtual dependency.")
|
||||||
|
|
||||||
# ensure patches is a list
|
# ensure patches is a list
|
||||||
@@ -520,13 +519,13 @@ def _depends_on(
|
|||||||
|
|
||||||
# this is where we actually add the dependency to this package
|
# this is where we actually add the dependency to this package
|
||||||
deps_by_name = pkg.dependencies.setdefault(when_spec, {})
|
deps_by_name = pkg.dependencies.setdefault(when_spec, {})
|
||||||
dependency = deps_by_name.get(dep_spec.name)
|
dependency = deps_by_name.get(spec.name)
|
||||||
|
|
||||||
if not dependency:
|
if not dependency:
|
||||||
dependency = Dependency(pkg, dep_spec, depflag=depflag)
|
dependency = Dependency(pkg, spec, depflag=depflag)
|
||||||
deps_by_name[dep_spec.name] = dependency
|
deps_by_name[spec.name] = dependency
|
||||||
else:
|
else:
|
||||||
dependency.spec.constrain(dep_spec, deps=False)
|
dependency.spec.constrain(spec, deps=False)
|
||||||
dependency.depflag |= depflag
|
dependency.depflag |= depflag
|
||||||
|
|
||||||
# apply patches to the dependency
|
# apply patches to the dependency
|
||||||
@@ -591,12 +590,13 @@ def depends_on(
|
|||||||
@see The section "Dependency specs" in the Spack Packaging Guide.
|
@see The section "Dependency specs" in the Spack Packaging Guide.
|
||||||
|
|
||||||
"""
|
"""
|
||||||
if spack.spec.Spec(spec).name in SUPPORTED_LANGUAGES:
|
dep_spec = spack.spec.Spec(spec)
|
||||||
|
if dep_spec.name in SUPPORTED_LANGUAGES:
|
||||||
assert type == "build", "languages must be of 'build' type"
|
assert type == "build", "languages must be of 'build' type"
|
||||||
return _language(lang_spec_str=spec, when=when)
|
return _language(lang_spec_str=spec, when=when)
|
||||||
|
|
||||||
def _execute_depends_on(pkg: "spack.package_base.PackageBase"):
|
def _execute_depends_on(pkg: "spack.package_base.PackageBase"):
|
||||||
_depends_on(pkg, spec, when=when, type=type, patches=patches)
|
_depends_on(pkg, dep_spec, when=when, type=type, patches=patches)
|
||||||
|
|
||||||
return _execute_depends_on
|
return _execute_depends_on
|
||||||
|
|
||||||
@@ -666,25 +666,24 @@ def extends(spec, when=None, type=("build", "run"), patches=None):
|
|||||||
|
|
||||||
keyword arguments can be passed to extends() so that extension
|
keyword arguments can be passed to extends() so that extension
|
||||||
packages can pass parameters to the extendee's extension
|
packages can pass parameters to the extendee's extension
|
||||||
mechanism.
|
mechanism."""
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
def _execute_extends(pkg):
|
def _execute_extends(pkg):
|
||||||
when_spec = _make_when_spec(when)
|
when_spec = _make_when_spec(when)
|
||||||
if not when_spec:
|
if not when_spec:
|
||||||
return
|
return
|
||||||
|
|
||||||
_depends_on(pkg, spec, when=when, type=type, patches=patches)
|
dep_spec = spack.spec.Spec(spec)
|
||||||
spec_obj = spack.spec.Spec(spec)
|
|
||||||
|
_depends_on(pkg, dep_spec, when=when, type=type, patches=patches)
|
||||||
|
|
||||||
# When extending python, also add a dependency on python-venv. This is done so that
|
# When extending python, also add a dependency on python-venv. This is done so that
|
||||||
# Spack environment views are Python virtual environments.
|
# Spack environment views are Python virtual environments.
|
||||||
if spec_obj.name == "python" and not pkg.name == "python-venv":
|
if dep_spec.name == "python" and not pkg.name == "python-venv":
|
||||||
_depends_on(pkg, "python-venv", when=when, type=("build", "run"))
|
_depends_on(pkg, spack.spec.Spec("python-venv"), when=when, type=("build", "run"))
|
||||||
|
|
||||||
# TODO: the values of the extendees dictionary are not used. Remove in next refactor.
|
# TODO: the values of the extendees dictionary are not used. Remove in next refactor.
|
||||||
pkg.extendees[spec_obj.name] = (spec_obj, None)
|
pkg.extendees[dep_spec.name] = (dep_spec, None)
|
||||||
|
|
||||||
return _execute_extends
|
return _execute_extends
|
||||||
|
|
||||||
|
|||||||
@@ -15,6 +15,7 @@
|
|||||||
|
|
||||||
import llnl.util.filesystem as fs
|
import llnl.util.filesystem as fs
|
||||||
import llnl.util.tty as tty
|
import llnl.util.tty as tty
|
||||||
|
from llnl.util.symlink import readlink
|
||||||
|
|
||||||
import spack.config
|
import spack.config
|
||||||
import spack.hash_types as ht
|
import spack.hash_types as ht
|
||||||
@@ -181,7 +182,7 @@ def deprecated_file_path(self, deprecated_spec, deprecator_spec=None):
|
|||||||
base_dir = (
|
base_dir = (
|
||||||
self.path_for_spec(deprecator_spec)
|
self.path_for_spec(deprecator_spec)
|
||||||
if deprecator_spec
|
if deprecator_spec
|
||||||
else os.readlink(deprecated_spec.prefix)
|
else readlink(deprecated_spec.prefix)
|
||||||
)
|
)
|
||||||
|
|
||||||
yaml_path = os.path.join(
|
yaml_path = os.path.join(
|
||||||
|
|||||||
@@ -34,6 +34,9 @@
|
|||||||
* ``spec``: a string representation of the abstract spec that was concretized
|
* ``spec``: a string representation of the abstract spec that was concretized
|
||||||
|
|
||||||
4. ``concrete_specs``: a dictionary containing the specs in the environment.
|
4. ``concrete_specs``: a dictionary containing the specs in the environment.
|
||||||
|
5. ``include_concrete`` (dictionary): an optional dictionary that includes the roots
|
||||||
|
and concrete specs from the included environments, keyed by the path to that
|
||||||
|
environment
|
||||||
|
|
||||||
Compatibility
|
Compatibility
|
||||||
-------------
|
-------------
|
||||||
@@ -50,26 +53,37 @@
|
|||||||
- ``v2``
|
- ``v2``
|
||||||
- ``v3``
|
- ``v3``
|
||||||
- ``v4``
|
- ``v4``
|
||||||
|
- ``v5``
|
||||||
* - ``v0.12:0.14``
|
* - ``v0.12:0.14``
|
||||||
- ✅
|
- ✅
|
||||||
-
|
-
|
||||||
-
|
-
|
||||||
-
|
-
|
||||||
|
-
|
||||||
* - ``v0.15:0.16``
|
* - ``v0.15:0.16``
|
||||||
- ✅
|
- ✅
|
||||||
- ✅
|
- ✅
|
||||||
-
|
-
|
||||||
-
|
-
|
||||||
|
-
|
||||||
* - ``v0.17``
|
* - ``v0.17``
|
||||||
- ✅
|
- ✅
|
||||||
- ✅
|
- ✅
|
||||||
- ✅
|
- ✅
|
||||||
-
|
-
|
||||||
|
-
|
||||||
* - ``v0.18:``
|
* - ``v0.18:``
|
||||||
- ✅
|
- ✅
|
||||||
- ✅
|
- ✅
|
||||||
- ✅
|
- ✅
|
||||||
- ✅
|
- ✅
|
||||||
|
-
|
||||||
|
* - ``v0.22:``
|
||||||
|
- ✅
|
||||||
|
- ✅
|
||||||
|
- ✅
|
||||||
|
- ✅
|
||||||
|
- ✅
|
||||||
|
|
||||||
Version 1
|
Version 1
|
||||||
---------
|
---------
|
||||||
@@ -334,6 +348,118 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
Version 5
|
||||||
|
---------
|
||||||
|
|
||||||
|
Version 5 doesn't change the top-level lockfile format, but an optional dictionary is
|
||||||
|
added. The dictionary has the ``root`` and ``concrete_specs`` of the included
|
||||||
|
environments, which are keyed by the path to that environment. Since this is optional
|
||||||
|
if the environment does not have any included environments ``include_concrete`` will
|
||||||
|
not be a part of the lockfile.
|
||||||
|
|
||||||
|
.. code-block:: json
|
||||||
|
|
||||||
|
{
|
||||||
|
"_meta": {
|
||||||
|
"file-type": "spack-lockfile",
|
||||||
|
"lockfile-version": 5,
|
||||||
|
"specfile-version": 3
|
||||||
|
},
|
||||||
|
"roots": [
|
||||||
|
{
|
||||||
|
"hash": "<dag_hash 1>",
|
||||||
|
"spec": "<abstract spec 1>"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"hash": "<dag_hash 2>",
|
||||||
|
"spec": "<abstract spec 2>"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"concrete_specs": {
|
||||||
|
"<dag_hash 1>": {
|
||||||
|
"... <spec dict attributes> ...": { },
|
||||||
|
"dependencies": [
|
||||||
|
{
|
||||||
|
"name": "depname_1",
|
||||||
|
"hash": "<dag_hash for depname_1>",
|
||||||
|
"type": ["build", "link"]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "depname_2",
|
||||||
|
"hash": "<dag_hash for depname_2>",
|
||||||
|
"type": ["build", "link"]
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"hash": "<dag_hash 1>",
|
||||||
|
},
|
||||||
|
"<daghash 2>": {
|
||||||
|
"... <spec dict attributes> ...": { },
|
||||||
|
"dependencies": [
|
||||||
|
{
|
||||||
|
"name": "depname_3",
|
||||||
|
"hash": "<dag_hash for depname_3>",
|
||||||
|
"type": ["build", "link"]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "depname_4",
|
||||||
|
"hash": "<dag_hash for depname_4>",
|
||||||
|
"type": ["build", "link"]
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"hash": "<dag_hash 2>"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"include_concrete": {
|
||||||
|
"<path to environment>": {
|
||||||
|
"roots": [
|
||||||
|
{
|
||||||
|
"hash": "<dag_hash 1>",
|
||||||
|
"spec": "<abstract spec 1>"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"hash": "<dag_hash 2>",
|
||||||
|
"spec": "<abstract spec 2>"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"concrete_specs": {
|
||||||
|
"<dag_hash 1>": {
|
||||||
|
"... <spec dict attributes> ...": { },
|
||||||
|
"dependencies": [
|
||||||
|
{
|
||||||
|
"name": "depname_1",
|
||||||
|
"hash": "<dag_hash for depname_1>",
|
||||||
|
"type": ["build", "link"]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "depname_2",
|
||||||
|
"hash": "<dag_hash for depname_2>",
|
||||||
|
"type": ["build", "link"]
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"hash": "<dag_hash 1>",
|
||||||
|
},
|
||||||
|
"<daghash 2>": {
|
||||||
|
"... <spec dict attributes> ...": { },
|
||||||
|
"dependencies": [
|
||||||
|
{
|
||||||
|
"name": "depname_3",
|
||||||
|
"hash": "<dag_hash for depname_3>",
|
||||||
|
"type": ["build", "link"]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"name": "depname_4",
|
||||||
|
"hash": "<dag_hash for depname_4>",
|
||||||
|
"type": ["build", "link"]
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"hash": "<dag_hash 2>"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
"""
|
"""
|
||||||
|
|
||||||
from .environment import (
|
from .environment import (
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -10,8 +10,9 @@
|
|||||||
import shutil
|
import shutil
|
||||||
import stat
|
import stat
|
||||||
import sys
|
import sys
|
||||||
from typing import Optional
|
from typing import Callable, Dict, Optional
|
||||||
|
|
||||||
|
from llnl.string import comma_or
|
||||||
from llnl.util import tty
|
from llnl.util import tty
|
||||||
from llnl.util.filesystem import (
|
from llnl.util.filesystem import (
|
||||||
mkdirp,
|
mkdirp,
|
||||||
@@ -49,19 +50,20 @@
|
|||||||
_projections_path = ".spack/projections.yaml"
|
_projections_path = ".spack/projections.yaml"
|
||||||
|
|
||||||
|
|
||||||
def view_symlink(src, dst, **kwargs):
|
LinkCallbackType = Callable[[str, str, "FilesystemView", Optional["spack.spec.Spec"]], None]
|
||||||
# keyword arguments are irrelevant
|
|
||||||
# here to fit required call signature
|
|
||||||
|
def view_symlink(src: str, dst: str, *args, **kwargs) -> None:
|
||||||
symlink(src, dst)
|
symlink(src, dst)
|
||||||
|
|
||||||
|
|
||||||
def view_hardlink(src, dst, **kwargs):
|
def view_hardlink(src: str, dst: str, *args, **kwargs) -> None:
|
||||||
# keyword arguments are irrelevant
|
|
||||||
# here to fit required call signature
|
|
||||||
os.link(src, dst)
|
os.link(src, dst)
|
||||||
|
|
||||||
|
|
||||||
def view_copy(src: str, dst: str, view, spec: Optional[spack.spec.Spec] = None):
|
def view_copy(
|
||||||
|
src: str, dst: str, view: "FilesystemView", spec: Optional["spack.spec.Spec"] = None
|
||||||
|
) -> None:
|
||||||
"""
|
"""
|
||||||
Copy a file from src to dst.
|
Copy a file from src to dst.
|
||||||
|
|
||||||
@@ -104,27 +106,40 @@ def view_copy(src: str, dst: str, view, spec: Optional[spack.spec.Spec] = None):
|
|||||||
tty.debug(f"Can't change the permissions for {dst}")
|
tty.debug(f"Can't change the permissions for {dst}")
|
||||||
|
|
||||||
|
|
||||||
def view_func_parser(parsed_name):
|
#: supported string values for `link_type` in an env, mapped to canonical values
|
||||||
# What method are we using for this view
|
_LINK_TYPES = {
|
||||||
if parsed_name in ("hardlink", "hard"):
|
"hardlink": "hardlink",
|
||||||
|
"hard": "hardlink",
|
||||||
|
"copy": "copy",
|
||||||
|
"relocate": "copy",
|
||||||
|
"add": "symlink",
|
||||||
|
"symlink": "symlink",
|
||||||
|
"soft": "symlink",
|
||||||
|
}
|
||||||
|
|
||||||
|
_VALID_LINK_TYPES = sorted(set(_LINK_TYPES.values()))
|
||||||
|
|
||||||
|
|
||||||
|
def canonicalize_link_type(link_type: str) -> str:
|
||||||
|
"""Return canonical"""
|
||||||
|
canonical = _LINK_TYPES.get(link_type)
|
||||||
|
if not canonical:
|
||||||
|
raise ValueError(
|
||||||
|
f"Invalid link type: '{link_type}. Must be one of {comma_or(_VALID_LINK_TYPES)}'"
|
||||||
|
)
|
||||||
|
return canonical
|
||||||
|
|
||||||
|
|
||||||
|
def function_for_link_type(link_type: str) -> LinkCallbackType:
|
||||||
|
link_type = canonicalize_link_type(link_type)
|
||||||
|
if link_type == "hardlink":
|
||||||
return view_hardlink
|
return view_hardlink
|
||||||
elif parsed_name in ("copy", "relocate"):
|
elif link_type == "symlink":
|
||||||
return view_copy
|
|
||||||
elif parsed_name in ("add", "symlink", "soft"):
|
|
||||||
return view_symlink
|
return view_symlink
|
||||||
else:
|
elif link_type == "copy":
|
||||||
raise ValueError(f"invalid link type for view: '{parsed_name}'")
|
return view_copy
|
||||||
|
|
||||||
|
assert False, "invalid link type" # need mypy Literal values
|
||||||
def inverse_view_func_parser(view_type):
|
|
||||||
# get string based on view type
|
|
||||||
if view_type is view_hardlink:
|
|
||||||
link_name = "hardlink"
|
|
||||||
elif view_type is view_copy:
|
|
||||||
link_name = "copy"
|
|
||||||
else:
|
|
||||||
link_name = "symlink"
|
|
||||||
return link_name
|
|
||||||
|
|
||||||
|
|
||||||
class FilesystemView:
|
class FilesystemView:
|
||||||
@@ -140,7 +155,16 @@ class FilesystemView:
|
|||||||
directory structure.
|
directory structure.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, root, layout, **kwargs):
|
def __init__(
|
||||||
|
self,
|
||||||
|
root: str,
|
||||||
|
layout: "spack.directory_layout.DirectoryLayout",
|
||||||
|
*,
|
||||||
|
projections: Optional[Dict] = None,
|
||||||
|
ignore_conflicts: bool = False,
|
||||||
|
verbose: bool = False,
|
||||||
|
link_type: str = "symlink",
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Initialize a filesystem view under the given `root` directory with
|
Initialize a filesystem view under the given `root` directory with
|
||||||
corresponding directory `layout`.
|
corresponding directory `layout`.
|
||||||
@@ -149,15 +173,14 @@ def __init__(self, root, layout, **kwargs):
|
|||||||
"""
|
"""
|
||||||
self._root = root
|
self._root = root
|
||||||
self.layout = layout
|
self.layout = layout
|
||||||
|
self.projections = {} if projections is None else projections
|
||||||
|
|
||||||
self.projections = kwargs.get("projections", {})
|
self.ignore_conflicts = ignore_conflicts
|
||||||
|
self.verbose = verbose
|
||||||
self.ignore_conflicts = kwargs.get("ignore_conflicts", False)
|
|
||||||
self.verbose = kwargs.get("verbose", False)
|
|
||||||
|
|
||||||
# Setup link function to include view
|
# Setup link function to include view
|
||||||
link_func = kwargs.get("link", view_symlink)
|
self.link_type = link_type
|
||||||
self.link = ft.partial(link_func, view=self)
|
self.link = ft.partial(function_for_link_type(link_type), view=self)
|
||||||
|
|
||||||
def add_specs(self, *specs, **kwargs):
|
def add_specs(self, *specs, **kwargs):
|
||||||
"""
|
"""
|
||||||
@@ -255,8 +278,24 @@ class YamlFilesystemView(FilesystemView):
|
|||||||
Filesystem view to work with a yaml based directory layout.
|
Filesystem view to work with a yaml based directory layout.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, root, layout, **kwargs):
|
def __init__(
|
||||||
super().__init__(root, layout, **kwargs)
|
self,
|
||||||
|
root: str,
|
||||||
|
layout: "spack.directory_layout.DirectoryLayout",
|
||||||
|
*,
|
||||||
|
projections: Optional[Dict] = None,
|
||||||
|
ignore_conflicts: bool = False,
|
||||||
|
verbose: bool = False,
|
||||||
|
link_type: str = "symlink",
|
||||||
|
):
|
||||||
|
super().__init__(
|
||||||
|
root,
|
||||||
|
layout,
|
||||||
|
projections=projections,
|
||||||
|
ignore_conflicts=ignore_conflicts,
|
||||||
|
verbose=verbose,
|
||||||
|
link_type=link_type,
|
||||||
|
)
|
||||||
|
|
||||||
# Super class gets projections from the kwargs
|
# Super class gets projections from the kwargs
|
||||||
# YAML specific to get projections from YAML file
|
# YAML specific to get projections from YAML file
|
||||||
@@ -638,9 +677,6 @@ class SimpleFilesystemView(FilesystemView):
|
|||||||
"""A simple and partial implementation of FilesystemView focused on performance and immutable
|
"""A simple and partial implementation of FilesystemView focused on performance and immutable
|
||||||
views, where specs cannot be removed after they were added."""
|
views, where specs cannot be removed after they were added."""
|
||||||
|
|
||||||
def __init__(self, root, layout, **kwargs):
|
|
||||||
super().__init__(root, layout, **kwargs)
|
|
||||||
|
|
||||||
def _sanity_check_view_projection(self, specs):
|
def _sanity_check_view_projection(self, specs):
|
||||||
"""A very common issue is that we end up with two specs of the same package, that project
|
"""A very common issue is that we end up with two specs of the same package, that project
|
||||||
to the same prefix. We want to catch that as early as possible and give a sensible error to
|
to the same prefix. We want to catch that as early as possible and give a sensible error to
|
||||||
|
|||||||
@@ -13,7 +13,6 @@
|
|||||||
import spack.config
|
import spack.config
|
||||||
import spack.relocate
|
import spack.relocate
|
||||||
from spack.util.elf import ElfParsingError, parse_elf
|
from spack.util.elf import ElfParsingError, parse_elf
|
||||||
from spack.util.executable import Executable
|
|
||||||
|
|
||||||
|
|
||||||
def is_shared_library_elf(filepath):
|
def is_shared_library_elf(filepath):
|
||||||
@@ -141,7 +140,7 @@ def post_install(spec, explicit=None):
|
|||||||
return
|
return
|
||||||
|
|
||||||
# Only enable on platforms using ELF.
|
# Only enable on platforms using ELF.
|
||||||
if not spec.satisfies("platform=linux") and not spec.satisfies("platform=cray"):
|
if not spec.satisfies("platform=linux"):
|
||||||
return
|
return
|
||||||
|
|
||||||
# Disable this hook when bootstrapping, to avoid recursion.
|
# Disable this hook when bootstrapping, to avoid recursion.
|
||||||
@@ -149,10 +148,9 @@ def post_install(spec, explicit=None):
|
|||||||
return
|
return
|
||||||
|
|
||||||
# Should failing to locate patchelf be a hard error?
|
# Should failing to locate patchelf be a hard error?
|
||||||
patchelf_path = spack.relocate._patchelf()
|
patchelf = spack.relocate._patchelf()
|
||||||
if not patchelf_path:
|
if not patchelf:
|
||||||
return
|
return
|
||||||
patchelf = Executable(patchelf_path)
|
|
||||||
|
|
||||||
fixes = find_and_patch_sonames(spec.prefix, spec.package.non_bindable_shared_objects, patchelf)
|
fixes = find_and_patch_sonames(spec.prefix, spec.package.non_bindable_shared_objects, patchelf)
|
||||||
|
|
||||||
|
|||||||
@@ -117,7 +117,7 @@ def post_install(spec, explicit=None):
|
|||||||
return
|
return
|
||||||
|
|
||||||
# Only enable on platforms using ELF.
|
# Only enable on platforms using ELF.
|
||||||
if not spec.satisfies("platform=linux") and not spec.satisfies("platform=cray"):
|
if not spec.satisfies("platform=linux"):
|
||||||
return
|
return
|
||||||
|
|
||||||
visit_directory_tree(spec.prefix, ElfFilesWithRPathVisitor())
|
visit_directory_tree(spec.prefix, ElfFilesWithRPathVisitor())
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user