-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Public MCTP Medium Physical Address to MCTP Endpoint D-Bus Object path #52
Comments
By DSP0236 v1.3.3 Figure 9 this statement isn't necessarily true:
The one EID can be assigned to multiple ports on a device, in which case the interface selection must be done by the kernel route table. Given that I think that we need to adjust your proposal a little. What thoughts do you have on accommodating multiple interfaces for a given EID? |
In that case, as my opinion, the |
So changing the address type isn't feasible without changing the object version. I think it's something we should try to avoid. Bear in mind that the object path is keyed by the EID, so using an array of addresses in the interface feels like it's shoe-horning the information into places its not designed to belong. Further, if one EID is mapped across multiple ports on a device, exposing this information in userspace isn't useful anyway as it shouldn't be the responsibility of userspace to make the choice. I think we need something more generic, possibly:
On this we can have one interface describing the network in which the device is participating, and another interface describing its physical ports, their properties, and the EID associated with each port. That way you can use the current known EID to look up the endpoint in mctpd, pull the UUID from the |
I think we need trims some character from UUID to make it compatible with D-Bus object path requirement. Right?
Do you mean the interface like below?
Do you have any idea about the Physical ports interface.
So the device has 2 ports:
|
There is no limit on object path length, but we would need to comply with the object name requirements, and format the UUID accordingly (mainly: no
Probably not. We would want to describe the connectivity to a device, rather than a set of network numbers (and then EIDs/hardware addresses elsewhere). We may be able to do that with an object path reference, but can discuss with Andrew on the design there.
I think we're missing part of the requirements here: how are you discovering those devices in the first place? If we just have the Medium Type available on the endpoint object, you should then have all you need to choose an endpoint, no? Overall, I think we need to make some overall design choices on whether we would "prefer" the split vs. unified EID approach here, and target our design decisions accordingly (of course, we'd still want to support both approaches regardless). I'll have a think about that and update shortly. |
But it's a matter of finding the set of EIDs associated with a specific device, right? If we don't have an index object such as something addressed by the UUID, then we have to iterate all the EID objects to correlate the UUIDs to find the faster interface. |
There are two ways the mctpd can discovery the device.
It is not enough because we can't map the medium type of one EID. The other things, we need to concern is the name of endpoint (Terminus name) which is required to create the PLDM sensors. Currently, the terminus name can come from PLDM "Overall system entity AUX name PDR" in
Yah, please things about that. In the current design, only |
Hi Jeremy, I check the Line 136 in dd944cf
peer has one MCTP physical address as current struct.Do you think we can add PhyAddress1 interface to current Endpoint object patch as current peer struct design?
|
Having a single physical addressing data on an endpoint object does not allow for the full set of addressing models proposed in DSP0236. As you say above, one endpoint may have multiple physical interfaces. It may also use one EID So: no, that proposal will not work here. I'm concerned about why we're attempting to expose the physical transport mechanism here; that's one of the primary implementation details that MCTP is aimed at abstracting! :) Is the motivation for this just to allow upper-layer applications (ie, pldmd) to select the best transport available? |
Yes. That is common use case of physical address D-Bus interface. The addition use case (At least in Ampere system) is matching the PLDM sensors of one PLDM terminus with one specific PCIe card which supports PLDM message. We can match MCTP physical address of PLDM-MCTP endpoint with PCIe address of that card from SMBios. |
ok, I think we can do much better with route metrics instead. That way, the application does not need to perform its own routing for bandwidth optimisation.
This sounds like an inventory management objective; wouldn't this have been handed at the time that the endpoint was first enumerated? |
In Ampere system, BMC is endpoint in PCIe-VDM, EID of PCIe devices are set by SoC. But those EIDs is not static. The PCIe address of PCIe devices is also enumerated at the SoC boot time ( PCIe address of one card is not static also). So to match the PLDM sensors with PCIe card we need to match the MCTP PCIe address with SMBios address every time the SoC boot up. There is not much OEM/OEM support PCIe-VDM, but I think at some point the others OEM/ODM will also face the similar issue as Ampere. Because the PCIe address of one PLDM-VDM endpoint is not static, in that case locate and naming the PLDM sensors of terminus will have trouble. |
ok, what does the "SoC" refer to here? the BMC, or something else? |
|
@ThuBaNguyen can you draw a diagram of your MCTP network(s), annotated with (TM)BOs? It would be helpful for me at least. Also, the discussion feels like it echoes some of that at #46 (comment) |
|
If S0 is the BO for the D1P1 link and the BMC is a TMBO, then it must be the case that the BMC is only the TMBO for the D1P2 link, as MCTP networks must be hierarchical. Further, there cannot be multiple TMBOs in a given network, so it must be the case that we have two networks:
The BMC is the TMBO for N1, and I presume S0 is the TMBO for N0. As there's no routing between networks (by definition in DSP0236 - we can set aside your statements about no cross-connection), the concept of using the physical transport medium of BMC:P1 and BMC:P2 to choose how to talk to S0 is an error of reasoning. They're incomparable as they're in separate networks. You must first choose the network you wish to use. In the context of your diagram, choosing the network uniquely determines the EID you must use with no regard for the physical transport, as there's only one relevant EID in each network. How you choose the networks is another matter, but I think consideration of physical transports is a misdirection. |
"there's no routing between networks" yes, in the mctp network there is not cross-connection between PCIe and SMBus.
I agree with this.
We still need know how to choose correct network (Chose N1 between N0 and N1 in Ampere case). |
I'm a bit wary of this statement with its "in the mctp network"; there shouldn't be one network for your scenario, you need two. Have you configured the MCTP stack with two?
Just to clarify, N1 is the SMBus link in my description above, N0 is the PCIe link.
Rather than over-generalise, can you exploit the host power state for your specific circumstances? |
Yes. I configured the MCTP stack with two. Similar as below.
Sorry,
In Ampere system, N1 (SMBus) is still available when N0 (PCIe) is up. So when BMC reboots with the host is on. N1 and N0 can be up at the same time. So as my opinion, we can't use the host power state in that case. |
If pldmd knows of the host power state and is aware of the network topology, the host power state can be used to select the network to use (regardless of the fact that both N0 and N1 are valid), right? Edit: perhaps this is getting a bit off-topic for the mctp repo, might be best to discuss it in OpenBMC forums? |
Above is the D-Bus property of MCTP network, Do you have any idea to chose which network to use?
|
We have similar use-case where MCTP applications protocol services (PLDM/SPDM/...) need to select a fastest transport where we have multiple paths to the same device. I don't think it matters whether or not the devices are on the same network. On our platforms, both paths to the device are active at all times (on the same power domain) and the BMC is the BO on both of them.. I see @jk-ozlabs 's comment above about using route metrics above, and while I agree that seems like the "right" way to approach this problem, how does one build these metrics in the first place? Wouldn't that entail running some dummy traffic for the first selection? Would it be OK if every endpoint object in mctpd exposes the underlying binding type as described in DSP0239? The intent being to use that in-lieu of port metrics as a "simpler" way to choose the right EID for a given device? |
That would happen as part of the initial route setup: we would assign a suitable metric at that time, likely based on a heuristic from the transport type. This would probably just be a fairly simple configuration on mctpd's side, just something like PCIe VDM transports get a metric of 1, SMBus transports get a metric of 2, ...
I'd rather this be based on a configured logic, rather than runtime measurement. |
The issue is that "underlying binding type" is not well defined. For example:
|
Ah, okay. I misunderstood your statement about metrics, then. So what you are proposing is rather than expose the binding type and have various users interpret them and duplicate that logic, we'd just expose a simple rank that apps can choose the "lowest" number from in case the same device (identified via its MCTP UUID) is reachable via multiple EIDs, did I get that right? |
Partially :) You're correct on not having every application decide on the best message routing internally. As you say, we would have mctpd set the priorities on specific transport types, so we only need to define that logic in one place. The metrics themselves would be attributes of a route, and are set on the kernel MCTP routing table. The kernel just chooses the route with the best metric for any given destination EID. So, if a PCIe VDM transport is available, it will be used in preference to SMBus. If that VDM transport goes down, SMBus becomes the route with the best metric, so we still have connectivity. Then, applications just send to a single EID of the destination endpoint, and the kernel handles the rest, based on the information that mctpd has provided in the initial route setup. This is based on a fairly standard setup of multiple IP links used in linux too; here my laptop ethernet link is used in preference to the wifi:
This behaviour does depend on adopting a MCTP network topology where each endpoint uses a single EID across all links though; otherwise we're back to a situation where each application needs to be making its own routing decisions, and also be aware of which EIDs are routable at any specific time. |
Ah! this all does make sense (esp. the parallels to IP networks). However, I am not sure we can impose the "single EID for a device" restriction. Would it hurt to have both mechanisms and leave it to the applications to choose in the case where EIDs are different? |
There can also be a case where the device supports only certain MCTP message types over certain transports (Ex. PLDM over PCIe and SPDM over I2C). I think any automatic route selection in the kernel would break such a case? |
"a MCTP network topology" includes BMC and multiple MCTP devices (CPU, PCIe cards) from different OEM/ODM, "single EID for a device" restriction requires the adopting in mctp-stack of those MCTP devices software as well.
This is my situation, in SMBus BMC is BO it will set EID 20 for S0. But in PCIe VDM, the S0 CPU firmware is BO and it will use different EID. I can't force the CPU firmware reuse EID 20. |
Yes, and where an endpoint only uses one EID.
I think that would still be okay. It is reasonable to expose the physical medium details for cases where the routing is managed externally (ie, by the application selecting an EID). (while I still think that's a suboptimal design, we can still accommodate it) However, we don't yet have a scheme for that defined which satisfies the issues I have outlined above. I suspect we will need a set of physical address details per endpoint, where that set may contain zero or more items. We'll also need to define the policy for physical addressing details on non-local endpoints. |
pldmd
can support multiple MCTP-Binding interface at the same time. It can also choose of interface with highest speed. Sopldmd
need to know the medium type to decide which medium interface should it use to communicate with EP when there are many medium interfaces.Type
inau.com.codeconstruct.MCTP.Endpoint1
with data type asbyte
to define the medium type asTable 2 – MCTP physical medium identifiers
in DSP0239 V1.3.0Address
inau.com.codeconstruct.MCTP.Endpoint1
with data type as string or byte array asTable 27 – Routing Table Entry format
in DSP0236 V1.3.0. Depends on the MCTP binding type the size of thisAddress
property can different. Maybe we can addAddressSize
property to identify the byte size of address.Table 27 – Routing Table Entry format
, theport number
should also be public to MCTP Endpoint D-Bus object path. When the EP is bridge we need toidentify a particular bus connection that the physical address for the entry is defined under
that bridge.The text was updated successfully, but these errors were encountered: