Skip to content

ceph: fix build with arrow 20#426609

Closed
ccicnce113424 wants to merge 1 commit intoNixOS:masterfrom
ccicnce113424:ceph-fix-build-with-arrow-20
Closed

ceph: fix build with arrow 20#426609
ccicnce113424 wants to merge 1 commit intoNixOS:masterfrom
ccicnce113424:ceph-fix-build-with-arrow-20

Conversation

@ccicnce113424
Copy link
Member

@ccicnce113424 ccicnce113424 commented Jul 19, 2025

Resolves #426401. Reverts #406483.

Things done

  • Built on platform:
    • x86_64-linux
    • aarch64-linux
    • x86_64-darwin
    • aarch64-darwin
  • Tested, as applicable:
  • Ran nixpkgs-review on this PR. See nixpkgs-review usage.
  • Tested basic functionality of all binary files, usually in ./result/bin/.
  • Nixpkgs Release Notes
    • Package update: when the change is major or breaking.
  • NixOS Release Notes
    • Module addition: when adding a new NixOS module.
    • Module update: when the change is significant.
  • Fits CONTRIBUTING.md, pkgs/README.md, maintainers/README.md and others READMEs.

Add a 👍 reaction to pull requests you find important.

@nixpkgs-ci nixpkgs-ci bot added 10.rebuild-linux: 1-10 This PR causes between 1 and 10 packages to rebuild on Linux. 10.rebuild-darwin: 0 This PR does not cause any packages to rebuild on Darwin. labels Jul 19, 2025
@nix-owners nix-owners bot requested review from adevress, benaryorg, johanot, krav and nh2 July 19, 2025 08:54
@Aleksanaa
Copy link
Member

nixpkgs-review result

Generated using nixpkgs-review-gha

Command: nixpkgs-review pr 426609

Logs: https://github.com/Aleksanaa/nixpkgs-review-gha/actions/runs/16399372226


x86_64-linux (sandbox = true)

❌ 4 packages failed to build:
  • ceph.passthru.tests.ceph-multi-node
  • ceph.passthru.tests.ceph-single-node
  • ceph.passthru.tests.ceph-single-node-bluestore
  • ceph.passthru.tests.ceph-single-node-bluestore-dmcrypt

Error logs: `x86_64-linux`
ceph.passthru.tests.ceph-multi-node
osd2 # [  936.546267] systemd[1]: Finished Cleanup of Temporary Directories.
osd1 # [  941.236740] systemd[1]: Starting Cleanup of Temporary Directories...
osd1 # [  941.545262] systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully.
osd1 # [  941.551772] systemd[1]: Finished Cleanup of Temporary Directories.
!!! Traceback (most recent call last):
!!!   File "", line 82, in 
!!!     monA.wait_until_succeeds("ceph -s | grep 'HEALTH_OK'")
!!! 
!!! RequestedAssertionFailed: action timed out after 900 tries with one-second pause in-between
cleanup
kill machine (pid 9)
qemu-system-x86_64: terminating on signal 15 from pid 6 (/nix/store/djck7mx6jad1w0yy6zings96dyxanls6-python3-3.13.5/bin/python3.13)
kill machine (pid 35)
qemu-system-x86_64: terminating on signal 15 from pid 6 (/nix/store/djck7mx6jad1w0yy6zings96dyxanls6-python3-3.13.5/bin/python3.13)
kill machine (pid 60)
qemu-system-x86_64: terminating on signal 15 from pid 6 (/nix/store/djck7mx6jad1w0yy6zings96dyxanls6-python3-3.13.5/bin/python3.13)
kill machine (pid 92)
qemu-system-x86_64: terminating on signal 15 from pid 6 (/nix/store/djck7mx6jad1w0yy6zings96dyxanls6-python3-3.13.5/bin/python3.13)
kill vlan (pid 7)
(finished: cleanup, in 0.05 seconds)
ceph.passthru.tests.ceph-single-node
monA # [  200.062364] ceph-osd[4683]: 2025-07-20T13:51:28.162+0000 7f6c624f76c0 -1 osd.0 0 waiting for initial osdmap
monA # [  200.092419] ceph-osd[4730]: 2025-07-20T13:51:28.191+0000 7fc3e10e86c0 -1 osd.2 6 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
monA # [  200.288763] ceph-osd[4683]: 2025-07-20T13:51:28.383+0000 7f6c5e2e86c0 -1 osd.0 6 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
monA: (finished: waiting for success: ceph osd stat | grep -e '3 osds: 3 up[^,]*, 3 in', in 33.76 seconds)
monA: waiting for success: ceph -s | grep 'mgr: a(active,'
monA: (finished: waiting for success: ceph -s | grep 'mgr: a(active,', in 1.85 seconds)
monA: waiting for success: ceph -s | grep 'HEALTH_OK'
monA # [  921.373771] systemd[1]: Starting Cleanup of Temporary Directories...
monA # [  921.910249] systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully.
monA # [  921.920742] systemd[1]: Finished Cleanup of Temporary Directories.
!!! Traceback (most recent call last):
!!!   File "", line 63, in 
!!!     monA.wait_until_succeeds("ceph -s | grep 'HEALTH_OK'")
!!! 
!!! RequestedAssertionFailed: action timed out after 900 tries with one-second pause in-between
cleanup
kill machine (pid 9)
qemu-system-x86_64: terminating on signal 15 from pid 6 (/nix/store/djck7mx6jad1w0yy6zings96dyxanls6-python3-3.13.5/bin/python3.13)
kill vlan (pid 7)
(finished: cleanup, in 0.00 seconds)
ceph.passthru.tests.ceph-single-node-bluestore
monA # [  203.697350] ceph-osd[4708]: 2025-07-20T13:51:30.642+0000 7f27ffee86c0 -1 osd.2 6 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
monA # [  203.800904] ceph-osd[4661]: 2025-07-20T13:51:30.750+0000 7f3a9e8e86c0 -1 osd.0 6 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
monA # [  203.817135] ceph-osd[4684]: 2025-07-20T13:51:30.766+0000 7f35b83f06c0 -1 osd.1 6 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
monA: (finished: waiting for success: ceph osd stat | grep -e '3 osds: 3 up[^,]*, 3 in', in 34.02 seconds)
monA: waiting for success: ceph -s | grep 'mgr: a(active,'
monA: (finished: waiting for success: ceph -s | grep 'mgr: a(active,', in 1.98 seconds)
monA: waiting for success: ceph -s | grep 'HEALTH_OK'
monA # [  903.105937] systemd[1]: Starting Cleanup of Temporary Directories...
monA # [  903.824273] systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully.
monA # [  903.836326] systemd[1]: Finished Cleanup of Temporary Directories.
!!! Traceback (most recent call last):
!!!   File "", line 63, in 
!!!     monA.wait_until_succeeds("ceph -s | grep 'HEALTH_OK'")
!!! 
!!! RequestedAssertionFailed: action timed out after 900 tries with one-second pause in-between
cleanup
kill machine (pid 9)
qemu-system-x86_64: terminating on signal 15 from pid 6 (/nix/store/djck7mx6jad1w0yy6zings96dyxanls6-python3-3.13.5/bin/python3.13)
kill vlan (pid 7)
(finished: cleanup, in 0.01 seconds)
ceph.passthru.tests.ceph-single-node-bluestore-dmcrypt
ceph # [  258.424253] ceph-mgr[5776]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve.
ceph # [  258.430655] ceph-mgr[5776]:   import numpy as np
ceph: (finished: waiting for success: ceph -s | grep 'mgr: ceph(active,', in 82.58 seconds)
ceph: waiting for success: ceph osd stat | grep -e '3 osds: 3 up[^,]*, 3 in'
ceph # [  263.149738] ceph-mgr[5776]: 2025-07-20T13:52:30.365+0000 7f222d1fd6c0 -1 mgr.server handle_report got status from non-daemon mon.0
ceph: (finished: waiting for success: ceph osd stat | grep -e '3 osds: 3 up[^,]*, 3 in', in 1.34 seconds)
ceph: waiting for success: ceph -s | grep 'HEALTH_OK'
ceph # [  939.396890] systemd[1]: Starting Cleanup of Temporary Directories...
ceph # [  940.057828] systemd[1]: systemd-tmpfiles-clean.service: Deactivated successfully.
ceph # [  940.065266] systemd[1]: Finished Cleanup of Temporary Directories.
!!! Traceback (most recent call last):
!!!   File "", line 71, in 
!!!     ceph.wait_until_succeeds("ceph -s | grep 'HEALTH_OK'")
!!! 
!!! RequestedAssertionFailed: action timed out after 900 tries with one-second pause in-between
cleanup
kill machine (pid 9)
qemu-system-x86_64: terminating on signal 15 from pid 6 (/nix/store/djck7mx6jad1w0yy6zings96dyxanls6-python3-3.13.5/bin/python3.13)
kill vlan (pid 7)
(finished: cleanup, in 0.01 seconds)

aarch64-linux (sandbox = true)

❌ 4 packages failed to build:
  • ceph.passthru.tests.ceph-multi-node
  • ceph.passthru.tests.ceph-single-node
  • ceph.passthru.tests.ceph-single-node-bluestore
  • ceph.passthru.tests.ceph-single-node-bluestore-dmcrypt

@ccicnce113424
Copy link
Member Author

I'm not familiar with ceph, I don't know the cause of the problem, I just want to try to fix the samba4Full build. I think it would be better to let someone more familiar with ceph handle this.

@Aleksanaa Aleksanaa reopened this Jul 20, 2025
@Aleksanaa
Copy link
Member

No, it may not be an issue with this. Maybe just false positive

@ccicnce113424
Copy link
Member Author

I reproduced the test failure and I have no idea why. I really need help.

@benaryorg
Copy link
Contributor

Digging a little into it (editing the test to show more output) I can see this here:

vm-test-run-basic-single-node-ceph-cluster-bluestore-dmcrypt>     health: HEALTH_WARN
vm-test-run-basic-single-node-ceph-cluster-bluestore-dmcrypt>             3 OSD(s) have broken BlueStore compression

Which seems definitely broken while also seeming like it could be related to arrow.
I'll dig further into this when I've got some more spoons.

@benaryorg
Copy link
Contributor

Just an FYI for anyone stumbling over this PR, most of the discussion is happening in #426401

@nixpkgs-ci nixpkgs-ci bot added the 2.status: merge conflict This PR has merge conflicts with the target branch label Jul 24, 2025
@ccicnce113424 ccicnce113424 force-pushed the ceph-fix-build-with-arrow-20 branch from 0e1f415 to 343ba27 Compare July 24, 2025 12:21
@nixpkgs-ci nixpkgs-ci bot removed the 2.status: merge conflict This PR has merge conflicts with the target branch label Jul 24, 2025
@djds
Copy link
Member

djds commented Jul 27, 2025

Thanks. I think this is working with reverted snappy #426401 (comment)

At least the things build.

@benaryorg
Copy link
Contributor

Also, yes, thanks for the PR from my side too.
Given that we're not forced to bump to arrow 20 yet, we'll likely stay on 19 just to make sure that we're not introducing breakage, especially since arrow seems to be tied to some cryptography code from what I remember.
If the next major release of Ceph hasn't baked in the patch for arrow 20 I might just cherry-pick this commit at that point tho ^^

@ccicnce113424
Copy link
Member Author

ccicnce113424 commented Jul 29, 2025

Superseded by #427905.

@djds
Copy link
Member

djds commented Jul 29, 2025

Superseded by #426609.

Isn't this a link to this PR?

@ccicnce113424
Copy link
Member Author

ccicnce113424 commented Jul 29, 2025

Isn't this a link to this PR?

@djds I copied the wrong PR number. It's #427905 actually.

djds added a commit to djds/nixpkgs that referenced this pull request Sep 17, 2025
Also patch s3select for arrow-cpp 20.

See: NixOS#426609
@ccicnce113424 ccicnce113424 deleted the ceph-fix-build-with-arrow-20 branch October 7, 2025 05:29
nh2 pushed a commit to nh2/nixpkgs that referenced this pull request Oct 26, 2025
Also patch s3select for arrow-cpp 20.

See: NixOS#426609

Also:

* Build with current `fmt`

Split (from the original larger contribution
NixOS#443671
) done by:
Niklas Hambüchen <[email protected]>
aviallon pushed a commit to aviallon/nixpkgs that referenced this pull request Nov 27, 2025
Also patch s3select for arrow-cpp 20.

See: NixOS#426609

Also:

* Build with current `fmt`

Split (from the original larger contribution
NixOS#443671
) done by:
Niklas Hambüchen <[email protected]>
numinit pushed a commit to numinit/nixpkgs that referenced this pull request Dec 1, 2025
Also patch s3select for arrow-cpp 20.

See: NixOS#426609

Also:

* Build with current `fmt`

Split (from the original larger contribution
NixOS#443671
) done by:
Niklas Hambüchen <[email protected]>
James-1701 pushed a commit to James-1701/nixpkgs that referenced this pull request Feb 27, 2026
Also patch s3select for arrow-cpp 20.

See: NixOS#426609

Also:

* Build with current `fmt`

Split (from the original larger contribution
NixOS#443671
) done by:
Niklas Hambüchen <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

10.rebuild-darwin: 0 This PR does not cause any packages to rebuild on Darwin. 10.rebuild-linux: 1-10 This PR causes between 1 and 10 packages to rebuild on Linux.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Build failure: ceph

4 participants