Date   

Re: DAOS in Docker

Lombardi, Johann
 

Hi there,

 

Did you load the uio_pci_generic module in the kernel as specific in the note?

 

Cheers,

Johann

 

From: <daos@daos.groups.io> on behalf of "helloworld@..." <helloworld@...>
Reply-To: "daos@daos.groups.io" <daos@daos.groups.io>
Date: Wednesday 12 August 2020 at 13:51
To: "daos@daos.groups.io" <daos@daos.groups.io>
Subject: [daos] DAOS in Docker

 

I'm configuring the DAOS in Docker with only-RAM emulation
When testing DAOS server, I met the /usr/bin/daos_io_server errors
How can I fix it ??

In details, "ERR  src/bio/bio_xstream.c:224 bio_spdk_env_init() Failed to init SPDK thread lib, DER_INVAL(-1003)"

I use the ~/daos/utils/config/examples/daos_server_local.yaml" as configuration file.

and 

scm_mount: /mnt/daos

scm_class: ram             

scm_size: 4                

                           

bdev_class: file           

bdev_size: 16              

bdev_list: [/tmp/daos-bdev]

---------------------------------------------------------------------
Intel Corporation SAS (French simplified joint stock company)
Registered headquarters: "Les Montalets"- 2, rue de Paris,
92196 Meudon Cedex, France
Registration Number:  302 456 199 R.C.S. NANTERRE
Capital: 4,572,000 Euros

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.


Avocado's upcoming LTS release

Cleber Rosa
 

Hi DAOS community,

Given that some of the DAOS testing[1] uses the Avocado testing
framework, i'd like to bring to your attention that we have an
upcoming 82.0 LTS release scheduled for Sept 7th[2].

For that release, we'd like to keep as much compatibility as possible,
and when not possible, allow for a smoother migration. 69.x LTS will
be maintained for another 6 months, after the release 82.0 LTS
release, but the sooner any issue is addressed, the better.

For that, we have an epic issue[3] in which we could use your help,
with:
* running the existing tests you have, with the most recent Avocado
version possible
* opening any issues[4] you encounter

This will feed into either bug fixes, or documentation on how to
migrate from 69.x LTS to 82.0 LTS.

In addition to this this, feel free to engage with us about how the
new Avocado features (and there's a lot of them) may be beneficial to
the Falco project.

Thanks!
- Cleber

--

[1] - https://github.com/daos-stack/daos/blob/master/src/tests/ftest/launch.py#L749
[2] - https://github.com/avocado-framework/avocado/milestone/8
[3] - https://github.com/avocado-framework/avocado/issues/4103
[4] - https://github.com/avocado-framework/avocado/issues/new/choose


DAOS in Docker

helloworld@...
 

I'm configuring the DAOS in Docker with only-RAM emulation
When testing DAOS server, I met the /usr/bin/daos_io_server errors
How can I fix it ??

In details, "ERR  src/bio/bio_xstream.c:224 bio_spdk_env_init() Failed to init SPDK thread lib, DER_INVAL(-1003)"

I use the ~/daos/utils/config/examples/daos_server_local.yaml" as configuration file.

and 

scm_mount: /mnt/daos

scm_class: ram             

scm_size: 4                

                           

bdev_class: file           

bdev_size: 16              

bdev_list: [/tmp/daos-bdev]


Re: RPM issues on CentOS 7.8

Murrell, Brian
 

On Fri, 2020-08-07 at 09:52 +0200, Steffen Christgau wrote:

Yeah, I noticed that. However, the documentation suggested the above
steps
(
https://daos-stack.github.io/admin/installation/#installing-daos-from-rpms
).
Using the a local repo appears to be more elegant. Maybe the
documentation can be updated with regards to that.
I've opened a ticket to resolve this discrepancy. Thanks for letting
us know.

Cheers,
b.


Re: RPM issues on CentOS 7.8

Steffen Christgau
 

On 04/08/2020 17.23, Murrell, Brian wrote:
On Thu, 2020-07-30 at 17:10 +0200, Steffen Christgau wrote:
Thus, I deleted the
package for the latter and also libabt-devel since it requires the
libabt package. After that I could install the other rpms.
Yes that is the correct thing to do. [...]

Just to be sure: Can the libabt package really be dropped or will it
be
missed by some component?
It can be dropped.

In addition: Are the developer packages really required to setup a
working system?
No. They are only needed if you want to build software with the DAOS
library.

I removed all '*devel*' rpms and did 'install *' which
succeeded as well.
That is fine.
Thanks for confirming those points.

FWIW, the tarball actually unpacks a fully functioning "repo-md"
repository that can be used with yum/dnf. You simply need to create
.repo files per the example in the README and then you can use commands
like "yum install daos-client" rather than just installing everything.
Yeah, I noticed that. However, the documentation suggested the above
steps
(https://daos-stack.github.io/admin/installation/#installing-daos-from-rpms).
Using the a local repo appears to be more elegant. Maybe the
documentation can be updated with regards to that.

Regards, Steffen


Re: DAOS on CentOS 8

Olivier, Jeffrey V
 

Hi Patrick,

 

Supported platforms are CentOS7, OpenSUSE 15.1, and Ubuntu 20.04

 

Most of the testing has been on CentOS7 so far though.

 

https://daos-stack.github.io/admin/installation/

 

I’d imagine CentOS8 is similar enough that it’s likely not a large porting effort.  We do have a ticket open to move to 8.

 

For my Fedora environment, I generally am first to find new compiler warnings and other such annoyances but I mostly work on VOS, the backend object storage for DAOS so don’t typically have the same requirements needed for a full stack.

 

-Jeff

 

From: <daos@daos.groups.io> on behalf of "Farrell, Patrick Arthur" <patrick.farrell@...>
Reply-To: "daos@daos.groups.io" <daos@daos.groups.io>
Date: Tuesday, August 4, 2020 at 12:21 PM
To: "daos@daos.groups.io" <daos@daos.groups.io>
Subject: Re: [daos] DAOS on CentOS 8

 

Jeff,

 

Currently, I'm finding many DAOS package dependencies are not in EPEL for CentOS 8.  Previously (last fall), I pushed through this - finding the dependencies from other, scattered sources, which obviously isn't ideal - and then I ran in to problems with the various Python packages which did not correctly handle the 'python' command being attached to python3.  Basically I ran in to a catch-22 where I couldn't get some packages to build if 'python' meant python3, and other issues with changing it back to python2.  It sounds like those are resolved if you're not having issues, so we can probably get CentOS 8 working if we push through the package issues.

 

Other than CentOS 7, does Intel have a preferred distribution to run on DAOS servers, or a list of those considered supported?

 

-Patrick


From: daos@daos.groups.io <daos@daos.groups.io> on behalf of Olivier, Jeffrey V <jeffrey.v.olivier@...>
Sent: Tuesday, August 4, 2020 11:36 AM
To: daos@daos.groups.io <daos@daos.groups.io>
Subject: Re: [daos] DAOS on CentOS 8

 

Hi Patrick,

 

I haven’t tried CentOS 8 but I’ve been using Fedora since version 27, now 32 and don’t recall issues encountered during the transition to python3 being the default.  Can you describe the issues you see ?

 

-Jeff

 

From: <daos@daos.groups.io> on behalf of "Farrell, Patrick Arthur" <patrick.farrell@...>
Reply-To: "daos@daos.groups.io" <daos@daos.groups.io>
Date: Wednesday, July 29, 2020 at 2:16 PM
To: "daos@daos.groups.io" <daos@daos.groups.io>
Subject: [daos] DAOS on CentOS 8

 

Good afternoon,

 

Has anyone been running DAOS on CentOS 8?  We tried this a while back, and had some significant issues related to 'python' referring to Python3 rather than Python2.  Before we try again, I was curious if anyone has been doing this & if there are any caveats or known problems.

 

Regards,

-Patrick


Re: DAOS on CentOS 8

Farrell, Patrick Arthur
 

Jeff,

Currently, I'm finding many DAOS package dependencies are not in EPEL for CentOS 8.  Previously (last fall), I pushed through this - finding the dependencies from other, scattered sources, which obviously isn't ideal - and then I ran in to problems with the various Python packages which did not correctly handle the 'python' command being attached to python3.  Basically I ran in to a catch-22 where I couldn't get some packages to build if 'python' meant python3, and other issues with changing it back to python2.  It sounds like those are resolved if you're not having issues, so we can probably get CentOS 8 working if we push through the package issues.

Other than CentOS 7, does Intel have a preferred distribution to run on DAOS servers, or a list of those considered supported?

-Patrick


From: daos@daos.groups.io <daos@daos.groups.io> on behalf of Olivier, Jeffrey V <jeffrey.v.olivier@...>
Sent: Tuesday, August 4, 2020 11:36 AM
To: daos@daos.groups.io <daos@daos.groups.io>
Subject: Re: [daos] DAOS on CentOS 8
 

Hi Patrick,

 

I haven’t tried CentOS 8 but I’ve been using Fedora since version 27, now 32 and don’t recall issues encountered during the transition to python3 being the default.  Can you describe the issues you see ?

 

-Jeff

 

From: <daos@daos.groups.io> on behalf of "Farrell, Patrick Arthur" <patrick.farrell@...>
Reply-To: "daos@daos.groups.io" <daos@daos.groups.io>
Date: Wednesday, July 29, 2020 at 2:16 PM
To: "daos@daos.groups.io" <daos@daos.groups.io>
Subject: [daos] DAOS on CentOS 8

 

Good afternoon,

 

Has anyone been running DAOS on CentOS 8?  We tried this a while back, and had some significant issues related to 'python' referring to Python3 rather than Python2.  Before we try again, I was curious if anyone has been doing this & if there are any caveats or known problems.

 

Regards,

-Patrick


Re: DAOS on CentOS 8

Olivier, Jeffrey V
 

Hi Patrick,

 

I haven’t tried CentOS 8 but I’ve been using Fedora since version 27, now 32 and don’t recall issues encountered during the transition to python3 being the default.  Can you describe the issues you see ?

 

-Jeff

 

From: <daos@daos.groups.io> on behalf of "Farrell, Patrick Arthur" <patrick.farrell@...>
Reply-To: "daos@daos.groups.io" <daos@daos.groups.io>
Date: Wednesday, July 29, 2020 at 2:16 PM
To: "daos@daos.groups.io" <daos@daos.groups.io>
Subject: [daos] DAOS on CentOS 8

 

Good afternoon,

 

Has anyone been running DAOS on CentOS 8?  We tried this a while back, and had some significant issues related to 'python' referring to Python3 rather than Python2.  Before we try again, I was curious if anyone has been doing this & if there are any caveats or known problems.

 

Regards,

-Patrick


Re: RPM issues on CentOS 7.8

Murrell, Brian
 

On Thu, 2020-07-30 at 17:10 +0200, Steffen Christgau wrote:
Thus, I deleted the
package for the latter and also libabt-devel since it requires the
libabt package. After that I could install the other rpms.
Yes that is the correct thing to do. The libabt* packages should not
be part of the tarball. We have an internal ticket to have this
rectified.

Just to be sure: Can the libabt package really be dropped or will it
be
missed by some component?
It can be dropped.

In addition: Are the developer packages really required to setup a
working system?
No. They are only needed if you want to build software with the DAOS
library.

I removed all '*devel*' rpms and did 'install *' which
succeeded as well.
That is fine.

FWIW, the tarball actually unpacks a fully functioning "repo-md"
repository that can be used with yum/dnf. You simply need to create
.repo files per the example in the README and then you can use commands
like "yum install daos-client" rather than just installing everything.

Looking forward for your responses.
Apologies for it being so late. I did actually respond to this last
week but I cannot find it. I must have wrote the message and forgot to
hit send, or got distracted or some such. Ahh. Just checked and it's
stuck in my Outbox. Must have been some kind of error trying to send
it.

Cheers,
b.


Re: RPM issues on CentOS 7.8

Kirvan, Scott
 

Steffen, 

I'll forward your questions to our packaging expert and have someone get back to you shortly.  One thing I noticed is that you are using 7.8 and for the DAOS 1.0 branch we did all of our testing on 7.7.  We moved to 7.8 with the 1.1/master branch that is under development.

Regards, Scott


From: daos@daos.groups.io <daos@daos.groups.io> on behalf of Steffen Christgau <christgau@...>
Sent: Thursday, July 30, 2020 9:10 AM
To: daos@daos.groups.io <daos@daos.groups.io>
Subject: [daos] RPM issues on CentOS 7.8
 
Hi everyone,

first, I really appreciate the availability of the RPMs for installing
DAOS. This really simplifies setting up a working system. I've tried
them and, in principle, succeeded to install DAOS. Nevertheless, I
haven't configured the system and thus haven't started it yet since
there are two minor issues I encountered and which I want to clearify
before continuing.

I'm using CentOS 7.8.2003, epel-release is already installed and up to
date. If I follow the admin guide and then do

yum install *

after unpacking the rpms I end up with the following conflicts:

Transaction check error:
  file /usr/lib64/libabt.so.0.0.0 conflicts between attempted installs
of libabt0-1.0rc1-2.el7.x86_64 and argobots-1.0rc1-5.el7.x86_64
  file /usr/include/abt.h conflicts between attempted installs of
argobots-devel-1.0rc1-5.el7.x86_64 and libabt-devel-1.0rc1-2.el7.x86_64
  file /usr/lib64/libabt.a conflicts between attempted installs of
argobots-devel-1.0rc1-5.el7.x86_64 and libabt-devel-1.0rc1-2.el7.x86_64

libabt and argobots provide the same content but different versions with
argobots appearing to be more recent than libabt. Thus, I deleted the
package for the latter and also libabt-devel since it requires the
libabt package. After that I could install the other rpms.

Just to be sure: Can the libabt package really be dropped or will it be
missed by some component?

In addition: Are the developer packages really required to setup a
working system? I removed all '*devel*' rpms and did 'install *' which
succeeded as well.

Looking forward for your responses.

Regards,

Steffen






Re: issues with NVMe drives from RPM installation

Dahringer, Richard
 

That worked!

 

Thanks Tom!

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of Nabarro, Tom
Sent: Thursday, July 30, 2020 12:11
To: daos@daos.groups.io
Subject: Re: [daos] issues with NVMe drives from RPM installation

 

Sounds like maybe metadata is out of sync, can you try removing /mnt/daos0/*, starting the server and then (on a separate tty) reformatting with "dmg storage format --reformat"?

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of Dahringer, Richard
Sent: Thursday, July 30, 2020 5:28 PM
To: daos@daos.groups.io
Subject: Re: [daos] issues with NVMe drives from RPM installation

 

Thanks Tom, that led me to this:


07/30-08:21:08.63 elfs13o01 DAOS[74504/74524] bio  INFO src/bio/bio_xstream.c:1049 bio_xsctxt_alloc() Initialize NVMe context, tgt_id:0, init_thread:(nil)

07/30-08:21:10.77 elfs13o01 DAOS[74504/74524] bio  ERR  src/bio/bio_xstream.c:877 init_blobstore_ctxt() Device list & device mapping is inconsistent

07/30-08:21:14.13 elfs13o01 DAOS[74504/74524] server ERR  src/iosrv/srv.c:452 dss_srv_handler() failed to init spdk context for xstream(2) rc:-1005

 

When I check for consistency, I see :

 

[root@elfs13o01 tmp]# daos_server storage scan

Scanning locally-attached storage...

ERROR: /usr/bin/daos_admin EAL: No free hugepages reported in hugepages-1048576kB

NVMe controllers and namespaces:

                PCI:0000:5e:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:0 Capacity:4.0 TB

                PCI:0000:5f:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:0 Capacity:4.0 TB

                PCI:0000:d8:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:1 Capacity:4.0 TB

                PCI:0000:d9:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:1 Capacity:4.0 TB

SCM Namespaces:

                Device:pmem0 Socket:0 Capacity:266 GB

                Device:pmem1 Socket:1 Capacity:266 GB

 

And the first line of the NVMe controllers listed is the drive I have in the configuration file (from below)

 

  bdev_class: nvme
  bdev_list: ["0000:5e:00.0"]

 

Is there another file somewhere that I need to set up?  I saw some documentation of ‘daos_nvme.conf’ which is automatically generated.  I added the second NVMe device on socket 0 to the configuration to test to see if that would change anything, but I have the same results.

 

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of Nabarro, Tom
Sent: Thursday, July 30, 2020 09:59
To: daos@daos.groups.io
Subject: Re: [daos] issues with NVMe drives from RPM installation

 

Hello Richard

 

"ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1”
indicates that there might be some useful information in the io_server log for the first instance, the default location as set in the server config file (log_file) is /tmp/server0.log. If nothing useful in there try increasing the log_mask to DEBUG.

Regards,

Tom Nabarro – HPC

M: +44 (0)7786 260986

Skype: tom.nabarro

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of richard.dahringer@...
Sent: Thursday, July 30, 2020 3:27 PM
To: daos@daos.groups.io
Subject: [daos] issues with NVMe drives from RPM installation

 

Hi all -
I'm trying to set up a proof of concept daos cluster, and it is proving to be tricky. The systems have 4 SCM 128G DIMMs, and 4 U.2 NVMe drives installed. I have installed all the RPMs from registrationcenter.intel.com, and have been able to set up the SCM devices, 'dmg -i' commands all seem to work.  When I add nvme drives to the configuration though, daos_server does not start - it does start when the nvme drives are not there. 

My daos_server.conf file:

name: daos_server
access_points: ['elfs13o01']
# port: 10001
provider: ofi+psm2
nr_hugepages: 4096
control_log_file: /tmp/daos_control.log
transport_config:
   allow_insecure: true

servers:
-
  targets: 1
  first_core: 0
  nr_xs_helpers: 0
  fabric_iface: hib0
  fabric_iface_port: 31416
  log_file: /tmp/daos_server.log

 

  env_vars:
  - DAOS_MD_CAP=1024
  - CRT_CTX_SHARE_ADDR=0
  - CRT_TIMEOUT=30
  - FI_SOCKETS_MAX_CONN_RETRY=1
  - FI_SOCKETS_CONN_TIMEOUT=2000

 

  # Storage definitions

 

  # When scm_class is set to ram, tmpfs will be used to emulate SCM.

  # The size of ram is specified by scm_size in GB units.

  scm_mount: /mnt/daos0  # map to -s /mnt/daos
  scm_class: dcpm
  scm_list: [/dev/pmem0]

  bdev_class: nvme
  bdev_list: ["0000:5e:00.0"]

The startup error:

[root@elfs13o01 ~]# daos_server -o daos_local.yml start
daos_server logging to file /tmp/daos_control.log
ERROR: /usr/bin/daos_admin EAL: No free hugepages reported in hugepages-1048576kB
DAOS Control Server (pid 73257) listening on 0.0.0.0:10001
Waiting for DAOS I/O Server instance storage to be ready...
SCM @ /mnt/daos0: 262 GB Total/247 GB Avail
Starting I/O server instance 0: /usr/bin/daos_io_server
daos_io_server:0 Using legacy core allocation algorithm
daos_io_server:0 Starting SPDK v19.04.1 / DPDK 19.02.0 initialization...
[ DPDK EAL parameters: daos -c 0x1 --pci-whitelist=0000:5e:00.0 --log-level=lib.eal:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk73258 --proc-type=auto ]
ERROR: daos_io_server:0 EAL: No free hugepages reported in hugepages-1048576kB
ERROR: /var/run/daos_server/daos_server.sock: failed to accept connection: accept unixpacket /var/run/daos_server/daos_server.sock: use of closed network connection
ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1

Can someone provide some pointers to what is going on? 

---------------------------------------------------------------------
Intel Corporation (UK) Limited
Registered No. 1134945 (England)
Registered Office: Pipers Way, Swindon SN3 1RJ
VAT No: 860 2173 47

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.

---------------------------------------------------------------------
Intel Corporation (UK) Limited
Registered No. 1134945 (England)
Registered Office: Pipers Way, Swindon SN3 1RJ
VAT No: 860 2173 47

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.


Re: issues with NVMe drives from RPM installation

Nabarro, Tom
 

Sounds like maybe metadata is out of sync, can you try removing /mnt/daos0/*, starting the server and then (on a separate tty) reformatting with "dmg storage format --reformat"?

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of Dahringer, Richard
Sent: Thursday, July 30, 2020 5:28 PM
To: daos@daos.groups.io
Subject: Re: [daos] issues with NVMe drives from RPM installation

 

Thanks Tom, that led me to this:


07/30-08:21:08.63 elfs13o01 DAOS[74504/74524] bio  INFO src/bio/bio_xstream.c:1049 bio_xsctxt_alloc() Initialize NVMe context, tgt_id:0, init_thread:(nil)

07/30-08:21:10.77 elfs13o01 DAOS[74504/74524] bio  ERR  src/bio/bio_xstream.c:877 init_blobstore_ctxt() Device list & device mapping is inconsistent

07/30-08:21:14.13 elfs13o01 DAOS[74504/74524] server ERR  src/iosrv/srv.c:452 dss_srv_handler() failed to init spdk context for xstream(2) rc:-1005

 

When I check for consistency, I see :

 

[root@elfs13o01 tmp]# daos_server storage scan

Scanning locally-attached storage...

ERROR: /usr/bin/daos_admin EAL: No free hugepages reported in hugepages-1048576kB

NVMe controllers and namespaces:

                PCI:0000:5e:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:0 Capacity:4.0 TB

                PCI:0000:5f:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:0 Capacity:4.0 TB

                PCI:0000:d8:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:1 Capacity:4.0 TB

                PCI:0000:d9:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:1 Capacity:4.0 TB

SCM Namespaces:

                Device:pmem0 Socket:0 Capacity:266 GB

                Device:pmem1 Socket:1 Capacity:266 GB

 

And the first line of the NVMe controllers listed is the drive I have in the configuration file (from below)

 

  bdev_class: nvme
  bdev_list: ["0000:5e:00.0"]

 

Is there another file somewhere that I need to set up?  I saw some documentation of ‘daos_nvme.conf’ which is automatically generated.  I added the second NVMe device on socket 0 to the configuration to test to see if that would change anything, but I have the same results.

 

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of Nabarro, Tom
Sent: Thursday, July 30, 2020 09:59
To: daos@daos.groups.io
Subject: Re: [daos] issues with NVMe drives from RPM installation

 

Hello Richard

 

"ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1”
indicates that there might be some useful information in the io_server log for the first instance, the default location as set in the server config file (log_file) is /tmp/server0.log. If nothing useful in there try increasing the log_mask to DEBUG.

Regards,

Tom Nabarro – HPC

M: +44 (0)7786 260986

Skype: tom.nabarro

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of richard.dahringer@...
Sent: Thursday, July 30, 2020 3:27 PM
To: daos@daos.groups.io
Subject: [daos] issues with NVMe drives from RPM installation

 

Hi all -
I'm trying to set up a proof of concept daos cluster, and it is proving to be tricky. The systems have 4 SCM 128G DIMMs, and 4 U.2 NVMe drives installed. I have installed all the RPMs from registrationcenter.intel.com, and have been able to set up the SCM devices, 'dmg -i' commands all seem to work.  When I add nvme drives to the configuration though, daos_server does not start - it does start when the nvme drives are not there. 

My daos_server.conf file:

name: daos_server
access_points: ['elfs13o01']
# port: 10001
provider: ofi+psm2
nr_hugepages: 4096
control_log_file: /tmp/daos_control.log
transport_config:
   allow_insecure: true

servers:
-
  targets: 1
  first_core: 0
  nr_xs_helpers: 0
  fabric_iface: hib0
  fabric_iface_port: 31416
  log_file: /tmp/daos_server.log

 

  env_vars:
  - DAOS_MD_CAP=1024
  - CRT_CTX_SHARE_ADDR=0
  - CRT_TIMEOUT=30
  - FI_SOCKETS_MAX_CONN_RETRY=1
  - FI_SOCKETS_CONN_TIMEOUT=2000

 

  # Storage definitions

 

  # When scm_class is set to ram, tmpfs will be used to emulate SCM.

  # The size of ram is specified by scm_size in GB units.

  scm_mount: /mnt/daos0  # map to -s /mnt/daos
  scm_class: dcpm
  scm_list: [/dev/pmem0]

  bdev_class: nvme
  bdev_list: ["0000:5e:00.0"]

The startup error:

[root@elfs13o01 ~]# daos_server -o daos_local.yml start
daos_server logging to file /tmp/daos_control.log
ERROR: /usr/bin/daos_admin EAL: No free hugepages reported in hugepages-1048576kB
DAOS Control Server (pid 73257) listening on 0.0.0.0:10001
Waiting for DAOS I/O Server instance storage to be ready...
SCM @ /mnt/daos0: 262 GB Total/247 GB Avail
Starting I/O server instance 0: /usr/bin/daos_io_server
daos_io_server:0 Using legacy core allocation algorithm
daos_io_server:0 Starting SPDK v19.04.1 / DPDK 19.02.0 initialization...
[ DPDK EAL parameters: daos -c 0x1 --pci-whitelist=0000:5e:00.0 --log-level=lib.eal:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk73258 --proc-type=auto ]
ERROR: daos_io_server:0 EAL: No free hugepages reported in hugepages-1048576kB
ERROR: /var/run/daos_server/daos_server.sock: failed to accept connection: accept unixpacket /var/run/daos_server/daos_server.sock: use of closed network connection
ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1

Can someone provide some pointers to what is going on? 

---------------------------------------------------------------------
Intel Corporation (UK) Limited
Registered No. 1134945 (England)
Registered Office: Pipers Way, Swindon SN3 1RJ
VAT No: 860 2173 47

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.

---------------------------------------------------------------------
Intel Corporation (UK) Limited
Registered No. 1134945 (England)
Registered Office: Pipers Way, Swindon SN3 1RJ
VAT No: 860 2173 47

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.


Re: issues with NVMe drives from RPM installation

Dahringer, Richard
 

Thanks Tom, that led me to this:


07/30-08:21:08.63 elfs13o01 DAOS[74504/74524] bio  INFO src/bio/bio_xstream.c:1049 bio_xsctxt_alloc() Initialize NVMe context, tgt_id:0, init_thread:(nil)

07/30-08:21:10.77 elfs13o01 DAOS[74504/74524] bio  ERR  src/bio/bio_xstream.c:877 init_blobstore_ctxt() Device list & device mapping is inconsistent

07/30-08:21:14.13 elfs13o01 DAOS[74504/74524] server ERR  src/iosrv/srv.c:452 dss_srv_handler() failed to init spdk context for xstream(2) rc:-1005

 

When I check for consistency, I see :

 

[root@elfs13o01 tmp]# daos_server storage scan

Scanning locally-attached storage...

ERROR: /usr/bin/daos_admin EAL: No free hugepages reported in hugepages-1048576kB

NVMe controllers and namespaces:

                PCI:0000:5e:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:0 Capacity:4.0 TB

                PCI:0000:5f:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:0 Capacity:4.0 TB

                PCI:0000:d8:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:1 Capacity:4.0 TB

                PCI:0000:d9:00.0 Model:INTEL SSDPE2KX040T8  FW:VDV10131 Socket:1 Capacity:4.0 TB

SCM Namespaces:

                Device:pmem0 Socket:0 Capacity:266 GB

                Device:pmem1 Socket:1 Capacity:266 GB

 

And the first line of the NVMe controllers listed is the drive I have in the configuration file (from below)

 

  bdev_class: nvme
  bdev_list: ["0000:5e:00.0"]

 

Is there another file somewhere that I need to set up?  I saw some documentation of ‘daos_nvme.conf’ which is automatically generated.  I added the second NVMe device on socket 0 to the configuration to test to see if that would change anything, but I have the same results.

 

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of Nabarro, Tom
Sent: Thursday, July 30, 2020 09:59
To: daos@daos.groups.io
Subject: Re: [daos] issues with NVMe drives from RPM installation

 

Hello Richard

 

"ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1”
indicates that there might be some useful information in the io_server log for the first instance, the default location as set in the server config file (log_file) is /tmp/server0.log. If nothing useful in there try increasing the log_mask to DEBUG.

Regards,

Tom Nabarro – HPC

M: +44 (0)7786 260986

Skype: tom.nabarro

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of richard.dahringer@...
Sent: Thursday, July 30, 2020 3:27 PM
To: daos@daos.groups.io
Subject: [daos] issues with NVMe drives from RPM installation

 

Hi all -
I'm trying to set up a proof of concept daos cluster, and it is proving to be tricky. The systems have 4 SCM 128G DIMMs, and 4 U.2 NVMe drives installed. I have installed all the RPMs from registrationcenter.intel.com, and have been able to set up the SCM devices, 'dmg -i' commands all seem to work.  When I add nvme drives to the configuration though, daos_server does not start - it does start when the nvme drives are not there. 

My daos_server.conf file:

name: daos_server
access_points: ['elfs13o01']
# port: 10001
provider: ofi+psm2
nr_hugepages: 4096
control_log_file: /tmp/daos_control.log
transport_config:
   allow_insecure: true

servers:
-
  targets: 1
  first_core: 0
  nr_xs_helpers: 0
  fabric_iface: hib0
  fabric_iface_port: 31416
  log_file: /tmp/daos_server.log

 

  env_vars:
  - DAOS_MD_CAP=1024
  - CRT_CTX_SHARE_ADDR=0
  - CRT_TIMEOUT=30
  - FI_SOCKETS_MAX_CONN_RETRY=1
  - FI_SOCKETS_CONN_TIMEOUT=2000

 

  # Storage definitions

 

  # When scm_class is set to ram, tmpfs will be used to emulate SCM.

  # The size of ram is specified by scm_size in GB units.

  scm_mount: /mnt/daos0  # map to -s /mnt/daos
  scm_class: dcpm
  scm_list: [/dev/pmem0]

  bdev_class: nvme
  bdev_list: ["0000:5e:00.0"]

The startup error:

[root@elfs13o01 ~]# daos_server -o daos_local.yml start
daos_server logging to file /tmp/daos_control.log
ERROR: /usr/bin/daos_admin EAL: No free hugepages reported in hugepages-1048576kB
DAOS Control Server (pid 73257) listening on 0.0.0.0:10001
Waiting for DAOS I/O Server instance storage to be ready...
SCM @ /mnt/daos0: 262 GB Total/247 GB Avail
Starting I/O server instance 0: /usr/bin/daos_io_server
daos_io_server:0 Using legacy core allocation algorithm
daos_io_server:0 Starting SPDK v19.04.1 / DPDK 19.02.0 initialization...
[ DPDK EAL parameters: daos -c 0x1 --pci-whitelist=0000:5e:00.0 --log-level=lib.eal:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk73258 --proc-type=auto ]
ERROR: daos_io_server:0 EAL: No free hugepages reported in hugepages-1048576kB
ERROR: /var/run/daos_server/daos_server.sock: failed to accept connection: accept unixpacket /var/run/daos_server/daos_server.sock: use of closed network connection
ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1

Can someone provide some pointers to what is going on? 

---------------------------------------------------------------------
Intel Corporation (UK) Limited
Registered No. 1134945 (England)
Registered Office: Pipers Way, Swindon SN3 1RJ
VAT No: 860 2173 47

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.


Re: issues with NVMe drives from RPM installation

Nabarro, Tom
 

Hello Richard

 

"ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1”
indicates that there might be some useful information in the io_server log for the first instance, the default location as set in the server config file (log_file) is /tmp/server0.log. If nothing useful in there try increasing the log_mask to DEBUG.

Regards,

Tom Nabarro – HPC

M: +44 (0)7786 260986

Skype: tom.nabarro

 

From: daos@daos.groups.io <daos@daos.groups.io> On Behalf Of richard.dahringer@...
Sent: Thursday, July 30, 2020 3:27 PM
To: daos@daos.groups.io
Subject: [daos] issues with NVMe drives from RPM installation

 

Hi all -
I'm trying to set up a proof of concept daos cluster, and it is proving to be tricky. The systems have 4 SCM 128G DIMMs, and 4 U.2 NVMe drives installed. I have installed all the RPMs from registrationcenter.intel.com, and have been able to set up the SCM devices, 'dmg -i' commands all seem to work.  When I add nvme drives to the configuration though, daos_server does not start - it does start when the nvme drives are not there. 

My daos_server.conf file:

name: daos_server
access_points: ['elfs13o01']
# port: 10001
provider: ofi+psm2
nr_hugepages: 4096
control_log_file: /tmp/daos_control.log
transport_config:
   allow_insecure: true

servers:
-
  targets: 1
  first_core: 0
  nr_xs_helpers: 0
  fabric_iface: hib0
  fabric_iface_port: 31416
  log_file: /tmp/daos_server.log

 

  env_vars:
  - DAOS_MD_CAP=1024
  - CRT_CTX_SHARE_ADDR=0
  - CRT_TIMEOUT=30
  - FI_SOCKETS_MAX_CONN_RETRY=1
  - FI_SOCKETS_CONN_TIMEOUT=2000

 

  # Storage definitions

 

  # When scm_class is set to ram, tmpfs will be used to emulate SCM.

  # The size of ram is specified by scm_size in GB units.

  scm_mount: /mnt/daos0  # map to -s /mnt/daos
  scm_class: dcpm
  scm_list: [/dev/pmem0]

  bdev_class: nvme
  bdev_list: ["0000:5e:00.0"]

The startup error:

[root@elfs13o01 ~]# daos_server -o daos_local.yml start
daos_server logging to file /tmp/daos_control.log
ERROR: /usr/bin/daos_admin EAL: No free hugepages reported in hugepages-1048576kB
DAOS Control Server (pid 73257) listening on 0.0.0.0:10001
Waiting for DAOS I/O Server instance storage to be ready...
SCM @ /mnt/daos0: 262 GB Total/247 GB Avail
Starting I/O server instance 0: /usr/bin/daos_io_server
daos_io_server:0 Using legacy core allocation algorithm
daos_io_server:0 Starting SPDK v19.04.1 / DPDK 19.02.0 initialization...
[ DPDK EAL parameters: daos -c 0x1 --pci-whitelist=0000:5e:00.0 --log-level=lib.eal:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk73258 --proc-type=auto ]
ERROR: daos_io_server:0 EAL: No free hugepages reported in hugepages-1048576kB
ERROR: /var/run/daos_server/daos_server.sock: failed to accept connection: accept unixpacket /var/run/daos_server/daos_server.sock: use of closed network connection
ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1

Can someone provide some pointers to what is going on? 

---------------------------------------------------------------------
Intel Corporation (UK) Limited
Registered No. 1134945 (England)
Registered Office: Pipers Way, Swindon SN3 1RJ
VAT No: 860 2173 47

This e-mail and any attachments may contain confidential material for
the sole use of the intended recipient(s). Any review or distribution
by others is strictly prohibited. If you are not the intended
recipient, please contact the sender and delete all copies.


RPM issues on CentOS 7.8

Steffen Christgau
 

Hi everyone,

first, I really appreciate the availability of the RPMs for installing
DAOS. This really simplifies setting up a working system. I've tried
them and, in principle, succeeded to install DAOS. Nevertheless, I
haven't configured the system and thus haven't started it yet since
there are two minor issues I encountered and which I want to clearify
before continuing.

I'm using CentOS 7.8.2003, epel-release is already installed and up to
date. If I follow the admin guide and then do

yum install *

after unpacking the rpms I end up with the following conflicts:

Transaction check error:
file /usr/lib64/libabt.so.0.0.0 conflicts between attempted installs
of libabt0-1.0rc1-2.el7.x86_64 and argobots-1.0rc1-5.el7.x86_64
file /usr/include/abt.h conflicts between attempted installs of
argobots-devel-1.0rc1-5.el7.x86_64 and libabt-devel-1.0rc1-2.el7.x86_64
file /usr/lib64/libabt.a conflicts between attempted installs of
argobots-devel-1.0rc1-5.el7.x86_64 and libabt-devel-1.0rc1-2.el7.x86_64

libabt and argobots provide the same content but different versions with
argobots appearing to be more recent than libabt. Thus, I deleted the
package for the latter and also libabt-devel since it requires the
libabt package. After that I could install the other rpms.

Just to be sure: Can the libabt package really be dropped or will it be
missed by some component?

In addition: Are the developer packages really required to setup a
working system? I removed all '*devel*' rpms and did 'install *' which
succeeded as well.

Looking forward for your responses.

Regards,

Steffen


Re: issues with NVMe drives from RPM installation

Farrell, Patrick Arthur
 

Richard,

There's nothing obviously wrong - to me, anyway - with your config, and no useful errors in the output.  You can check the logs in /tmp/daos*.log (There will be multiple files), they should contain more information.  You could also turn on debug before you start the server to possibly get more info - described in the manual https://daos-stack.github.io/admin/troubleshooting/

Also, if you have not, you can check your drives are visible to DAOS and can be prepared as expected with the daos_server storage commands, scan and prepare, detailed here:
https://daos-stack.github.io/admin/deployment/

That details how to run them for SCM, look at the command help for how to run them for NVMe devices.  (You'll want to select NVMe only or it may ask you to reboot to set up your SCM goals, which you've obviously already done.)

Regards,
-Patrick


From: daos@daos.groups.io <daos@daos.groups.io> on behalf of richard.dahringer@... <richard.dahringer@...>
Sent: Thursday, July 30, 2020 9:27 AM
To: daos@daos.groups.io <daos@daos.groups.io>
Subject: [daos] issues with NVMe drives from RPM installation
 

Hi all -
I'm trying to set up a proof of concept daos cluster, and it is proving to be tricky. The systems have 4 SCM 128G DIMMs, and 4 U.2 NVMe drives installed. I have installed all the RPMs from registrationcenter.intel.com, and have been able to set up the SCM devices, 'dmg -i' commands all seem to work.  When I add nvme drives to the configuration though, daos_server does not start - it does start when the nvme drives are not there. 

My daos_server.conf file:

name: daos_server
access_points: ['elfs13o01']
# port: 10001
provider: ofi+psm2
nr_hugepages: 4096
control_log_file: /tmp/daos_control.log
transport_config:
   allow_insecure: true

servers:
-
  targets: 1
  first_core: 0
  nr_xs_helpers: 0
  fabric_iface: hib0
  fabric_iface_port: 31416
  log_file: /tmp/daos_server.log

 

  env_vars:
  - DAOS_MD_CAP=1024
  - CRT_CTX_SHARE_ADDR=0
  - CRT_TIMEOUT=30
  - FI_SOCKETS_MAX_CONN_RETRY=1
  - FI_SOCKETS_CONN_TIMEOUT=2000

 

  # Storage definitions

 

  # When scm_class is set to ram, tmpfs will be used to emulate SCM.

  # The size of ram is specified by scm_size in GB units.

  scm_mount: /mnt/daos0  # map to -s /mnt/daos
  scm_class: dcpm
  scm_list: [/dev/pmem0]

  bdev_class: nvme
  bdev_list: ["0000:5e:00.0"]

The startup error:

[root@elfs13o01 ~]# daos_server -o daos_local.yml start
daos_server logging to file /tmp/daos_control.log
ERROR: /usr/bin/daos_admin EAL: No free hugepages reported in hugepages-1048576kB
DAOS Control Server (pid 73257) listening on 0.0.0.0:10001
Waiting for DAOS I/O Server instance storage to be ready...
SCM @ /mnt/daos0: 262 GB Total/247 GB Avail
Starting I/O server instance 0: /usr/bin/daos_io_server
daos_io_server:0 Using legacy core allocation algorithm
daos_io_server:0 Starting SPDK v19.04.1 / DPDK 19.02.0 initialization...
[ DPDK EAL parameters: daos -c 0x1 --pci-whitelist=0000:5e:00.0 --log-level=lib.eal:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk73258 --proc-type=auto ]
ERROR: daos_io_server:0 EAL: No free hugepages reported in hugepages-1048576kB
ERROR: /var/run/daos_server/daos_server.sock: failed to accept connection: accept unixpacket /var/run/daos_server/daos_server.sock: use of closed network connection
ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1

Can someone provide some pointers to what is going on? 


issues with NVMe drives from RPM installation

Dahringer, Richard
 

Hi all -
I'm trying to set up a proof of concept daos cluster, and it is proving to be tricky. The systems have 4 SCM 128G DIMMs, and 4 U.2 NVMe drives installed. I have installed all the RPMs from registrationcenter.intel.com, and have been able to set up the SCM devices, 'dmg -i' commands all seem to work.  When I add nvme drives to the configuration though, daos_server does not start - it does start when the nvme drives are not there. 

My daos_server.conf file:

name: daos_server
access_points: ['elfs13o01']
# port: 10001
provider: ofi+psm2
nr_hugepages: 4096
control_log_file: /tmp/daos_control.log
transport_config:
   allow_insecure: true

servers:
-
  targets: 1
  first_core: 0
  nr_xs_helpers: 0
  fabric_iface: hib0
  fabric_iface_port: 31416
  log_file: /tmp/daos_server.log

 

  env_vars:
  - DAOS_MD_CAP=1024
  - CRT_CTX_SHARE_ADDR=0
  - CRT_TIMEOUT=30
  - FI_SOCKETS_MAX_CONN_RETRY=1
  - FI_SOCKETS_CONN_TIMEOUT=2000

 

  # Storage definitions

 

  # When scm_class is set to ram, tmpfs will be used to emulate SCM.

  # The size of ram is specified by scm_size in GB units.

  scm_mount: /mnt/daos0  # map to -s /mnt/daos
  scm_class: dcpm
  scm_list: [/dev/pmem0]

  bdev_class: nvme
  bdev_list: ["0000:5e:00.0"]

The startup error:

[root@elfs13o01 ~]# daos_server -o daos_local.yml start
daos_server logging to file /tmp/daos_control.log
ERROR: /usr/bin/daos_admin EAL: No free hugepages reported in hugepages-1048576kB
DAOS Control Server (pid 73257) listening on 0.0.0.0:10001
Waiting for DAOS I/O Server instance storage to be ready...
SCM @ /mnt/daos0: 262 GB Total/247 GB Avail
Starting I/O server instance 0: /usr/bin/daos_io_server
daos_io_server:0 Using legacy core allocation algorithm
daos_io_server:0 Starting SPDK v19.04.1 / DPDK 19.02.0 initialization...
[ DPDK EAL parameters: daos -c 0x1 --pci-whitelist=0000:5e:00.0 --log-level=lib.eal:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk73258 --proc-type=auto ]
ERROR: daos_io_server:0 EAL: No free hugepages reported in hugepages-1048576kB
ERROR: /var/run/daos_server/daos_server.sock: failed to accept connection: accept unixpacket /var/run/daos_server/daos_server.sock: use of closed network connection
ERROR: DAOS I/O Server exited with error: /usr/bin/daos_io_server (instance 0) exited: exit status 1

Can someone provide some pointers to what is going on? 


DAOS on CentOS 8

Farrell, Patrick Arthur
 

Good afternoon,

Has anyone been running DAOS on CentOS 8?  We tried this a while back, and had some significant issues related to 'python' referring to Python3 rather than Python2.  Before we try again, I was curious if anyone has been doing this & if there are any caveats or known problems.

Regards,
-Patrick


Re: DAOS ior module has compilation errors

Chaarawi, Mohamad
 

Hi Kevan,

 

Yes I missed this in the latest fixes to update the DAOS & DFS drivers to the new IOR backend API changes.

This PR fixes it:

https://github.com/hpc/ior/pull/244

 

hopefully should land quickly.

 

Thanks,

Mohamad

 

From: <daos@daos.groups.io> on behalf of Kevan Rehm <kevan.rehm@...>
Reply-To: "daos@daos.groups.io" <daos@daos.groups.io>
Date: Wednesday, July 22, 2020 at 9:55 AM
To: "daos@daos.groups.io" <daos@daos.groups.io>
Subject: [daos] DAOS ior module has compilation errors

 

Apologies if this is old news, but the upstream ior master branch src/aiori-DAOS.c file does not compile cleanly, the signature apparently has changed for .get_file_size, the MPI_Comm parameter has been removed.

 

Regards, Kevan

 

  CC       libaiori_a-aiori-DUMMY.o

  CC       libaiori_a-aiori-MPIIO.o

  CC       libaiori_a-aiori-MMAP.o

  CC       libaiori_a-aiori-POSIX.o

  CC       libaiori_a-aiori-DAOS.o

aiori-DAOS.c:109:9: warning: initialization from incompatible pointer type [enabled by default]

         .get_file_size = DAOS_GetFileSize,

         ^

aiori-DAOS.c:109:9: warning: (near initialization for ‘daos_aiori.get_file_size’) [enabled by default]

  CC       libaiori_a-aiori-DFS.o

  AR       libaiori.a

  CC       ior-ior-main.o

  CC       ior-aiori.o

  CC       ior-aiori-DUMMY.o

  CC       ior-aiori-MPIIO.o

  CC       ior-aiori-MMAP.o

  CC       ior-aiori-POSIX.o

  CC       ior-aiori-DAOS.o

aiori-DAOS.c:109:9: warning: initialization from incompatible pointer type [enabled by default]

         .get_file_size = DAOS_GetFileSize,

         ^

aiori-DAOS.c:109:9: warning: (near initialization for ‘daos_aiori.get_file_size’) [enabled by default]

  CC       ior-aiori-DFS.o

  CCLD     ior

  CC       mdtest-mdtest-main.o

  CC       mdtest-aiori.o

  CC       mdtest-aiori-DUMMY.o

  CC       mdtest-aiori-MPIIO.o

  CC       mdtest-aiori-MMAP.o

  CC       mdtest-aiori-POSIX.o

  CC       mdtest-aiori-DAOS.o

aiori-DAOS.c:109:9: warning: initialization from incompatible pointer type [enabled by default]

         .get_file_size = DAOS_GetFileSize,

         ^

aiori-DAOS.c:109:9: warning: (near initialization for ‘daos_aiori.get_file_size’) [enabled by default]

  CC       mdtest-aiori-DFS.o

  CCLD     mdtest

  CC       test/lib.o


DAOS ior module has compilation errors

Kevan Rehm
 

Apologies if this is old news, but the upstream ior master branch src/aiori-DAOS.c file does not compile cleanly, the signature apparently has changed for .get_file_size, the MPI_Comm parameter has been removed.

 

Regards, Kevan

 

  CC       libaiori_a-aiori-DUMMY.o

  CC       libaiori_a-aiori-MPIIO.o

  CC       libaiori_a-aiori-MMAP.o

  CC       libaiori_a-aiori-POSIX.o

  CC       libaiori_a-aiori-DAOS.o

aiori-DAOS.c:109:9: warning: initialization from incompatible pointer type [enabled by default]

         .get_file_size = DAOS_GetFileSize,

         ^

aiori-DAOS.c:109:9: warning: (near initialization for ‘daos_aiori.get_file_size’) [enabled by default]

  CC       libaiori_a-aiori-DFS.o

  AR       libaiori.a

  CC       ior-ior-main.o

  CC       ior-aiori.o

  CC       ior-aiori-DUMMY.o

  CC       ior-aiori-MPIIO.o

  CC       ior-aiori-MMAP.o

  CC       ior-aiori-POSIX.o

  CC       ior-aiori-DAOS.o

aiori-DAOS.c:109:9: warning: initialization from incompatible pointer type [enabled by default]

         .get_file_size = DAOS_GetFileSize,

         ^

aiori-DAOS.c:109:9: warning: (near initialization for ‘daos_aiori.get_file_size’) [enabled by default]

  CC       ior-aiori-DFS.o

  CCLD     ior

  CC       mdtest-mdtest-main.o

  CC       mdtest-aiori.o

  CC       mdtest-aiori-DUMMY.o

  CC       mdtest-aiori-MPIIO.o

  CC       mdtest-aiori-MMAP.o

  CC       mdtest-aiori-POSIX.o

  CC       mdtest-aiori-DAOS.o

aiori-DAOS.c:109:9: warning: initialization from incompatible pointer type [enabled by default]

         .get_file_size = DAOS_GetFileSize,

         ^

aiori-DAOS.c:109:9: warning: (near initialization for ‘daos_aiori.get_file_size’) [enabled by default]

  CC       mdtest-aiori-DFS.o

  CCLD     mdtest

  CC       test/lib.o