Add web post update for 2025 #176

Manually merged
rarias merged 1 commits from post-update-2025 into master 2025-09-29 18:04:37 +02:00
Owner

Use hugo server -Ds web/ to test. Anything else I missed?

Use `hugo server -Ds web/` to test. Anything else I missed?
rarias added 1 commit 2025-09-26 14:58:34 +02:00
rarias requested review from abonerib 2025-09-26 14:58:34 +02:00
abonerib reviewed 2025-09-26 17:20:06 +02:00
abonerib left a comment
Collaborator

Sorry if I have gone overboard with the suggestions.

Maybe we can also mention that now we serve a binary cache, so people don't have to rebuild bscpkgs.

Towards the end, there are some sos that could be removed to make the text more fluid.

Sorry if I have gone overboard with the suggestions. Maybe we can also mention that now we serve a binary cache, so people don't have to rebuild bscpkgs. Towards the end, there are some `so`s that could be removed to make the text more fluid.
@ -0,0 +13,4 @@
We have a new [fox machine](/fox), with two AMD Genoa 9684X CPUs and two NVIDIA
RTX4000 GPUs. During the last months we have been doing some tests and it seems
that most of the components work well. We have configured CUDA to use the NVIDIA
GPUs as well as AMD uProf to trace performance and energy counters from the
Collaborator

I feel like a comma after GPUs would make things more clear.

I feel like a comma after GPUs would make things more clear.
rarias marked this conversation as resolved
@ -0,0 +20,4 @@
We have upgraded the operating system on the login node to NixOS, which now runs
Linux 6.15.6. During the upgrade, we have detected a problem with the RAID
controller that caused a catastrophic failure that prevented the BIOS from
Collaborator

the second that could be changed to which to avoid repetition

the second `that` could be changed to `which` to avoid repetition
rarias marked this conversation as resolved
@ -0,0 +23,4 @@
controller that caused a catastrophic failure that prevented the BIOS from
starting.
The `/` and `/home` partitions sit on a RAID 5 governed by a RAID hardware
Collaborator

We are still talking about the RAID controller, so splitting the paragraph is a bit confusing (Unless we change the section header to problems with the raid controller).

We are still talking about the RAID controller, so splitting the paragraph is a bit confusing (Unless we change the section header to problems with the raid controller).
Author
Owner

I rewrote it to make it more clear.

I rewrote it to make it more clear.
rarias marked this conversation as resolved
@ -0,0 +27,4 @@
controller, however it was unable to boot properly before handling
the control over to the BIOS. After a long debugging session, we detected that
the flash memory that stores the firmware of the hardware controller was likely
to be the issue, as
Collaborator

as since memory (as is grammatically correct, but using it here reads as while: e.g. as/while memory cells lose charge they do X). (https://writinglawtutors.com/dont-use-as-to-mean-because/)

~as~ since memory (as is grammatically correct, but using it here reads as while: e.g. `as/while memory cells lose charge they do X`). (https://writinglawtutors.com/dont-use-as-to-mean-because/)
rarias marked this conversation as resolved
@ -0,0 +29,4 @@
the flash memory that stores the firmware of the hardware controller was likely
to be the issue, as
[memory cells](https://en.wikipedia.org/wiki/Flash_memory#Principles_of_operation)
may lose charge over time and can end up corrupting the content. So we flashed
Collaborator

I would drop the first So since it's a crutch.

I would drop the first `So` since it's a crutch.
rarias marked this conversation as resolved
@ -0,0 +33,4 @@
the latest firmware so the memory cells are charged again with the new bits and
that fixed the problem. Hopefully we will be able to use it for some more years.
The SLURM server has been moved to apex, so now you can allocate your jobs from
Collaborator

The rest of the sos are fine, although they are a bit repetitve.

The rest of the `so`s are fine, although they are a bit repetitve.
rarias marked this conversation as resolved
@ -0,0 +36,4 @@
The SLURM server has been moved to apex, so now you can allocate your jobs from
there, including the new fox machine.
### Translated machines to BSC building
Collaborator

Transferred / Migrated

Transferred / Migrated
rarias marked this conversation as resolved
@ -0,0 +38,4 @@
### Translated machines to BSC building
The server room had a temperature issue that affected our machines since the end
Collaborator

had been affecting

had been affecting
rarias marked this conversation as resolved
@ -0,0 +44,4 @@
![Room temperature](temp.png)
Since then, we have moved the cluster to BSC premises, where now rests at a
Collaborator

where it now

where **it** now
rarias marked this conversation as resolved
rarias force-pushed post-update-2025 from e8eb47c9b8 to c1e042be96 2025-09-29 17:44:22 +02:00 Compare
Author
Owner

Maybe we can also mention that now we serve a binary cache, so people don't have to rebuild bscpkgs.

I'll wait until we have checked that we are not exposing anything that should not be in the cache. I would also like to test it on non-nixos machines to see how that would work.

> Maybe we can also mention that now we serve a binary cache, so people don't have to rebuild bscpkgs. I'll wait until we have checked that we are not exposing anything that should not be in the cache. I would also like to test it on non-nixos machines to see how that would work.
rarias removed review request for abonerib 2025-09-29 17:58:15 +02:00
rarias requested review from abonerib 2025-09-29 17:58:19 +02:00
abonerib approved these changes 2025-09-29 17:59:56 +02:00
rarias force-pushed post-update-2025 from c1e042be96 to c441178910 2025-09-29 18:02:48 +02:00 Compare
rarias manually merged commit c441178910 into master 2025-09-29 18:04:37 +02:00
Sign in to join this conversation.
No Reviewers
2 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: rarias/jungle#176
No description provided.