This post is older than 2 years and might not be relevant anymore
More Info: Consider searching for newer posts

BLE connection stability

Hello,

I would appreciate some guidance about BLE connection stability. Our product is a system of 3 peripherals (custom board with Laird's BL652 module, containing nRF52832) all simultaneously connected to our iOS app (central). The peripheral application is written with Nordic SDK 15.3 plus S112 v6.1.1, based initially on the ble_app_uart example. Our iOS app uses Apple's Core Bluetooth framework and currently supports iOS 9.3 or newer.

Each peripheral has a sensor and uses GPIOTE, PPI and TIMERs to timestamp events and send to the connected central via NUS. Events occur randomly, sometimes with a minute or two between events, and sometimes multiple events occur in quick succession. The clocks (TIMER2) on the three peripherals are synchronized using the Timeslot API based on a simplified version of this Nordic blog. The timeslot length is 1000 microseconds.

For higher accuracy, our peripheral app requests the high frequency clock (sd_clock_hfclk_request()), which I understand sets the clock source to XTAL. The BL652 has an integrated high accuracy 32 MHz (±10 ppm) crystal oscillator.

This is a mobile system that is set up with each peripheral at most 150 feet apart (open space, line-of-sight). Each peripheral is powered by 2 AA batteries. For testing, we set up three systems to test concurrent use of multiple 3-peripheral systems connected to multiple iOS devices. Each of the three systems (each comprising of 3 peripherals) uses a different radio address for the time sync using the Timeslot API.

Connection parameters for the peripherals are:

#define MIN_CONN_INTERVAL MSEC_TO_UNITS(30, UNIT_1_25_MS)
#define MAX_CONN_INTERVAL MSEC_TO_UNITS(75, UNIT_1_25_MS)
#define SLAVE_LATENCY 0
#define CONN_SUP_TIMEOUT MSEC_TO_UNITS(4000, UNIT_10_MS)
#define FIRST_CONN_PARAMS_UPDATE_DELAY APP_TIMER_TICKS(5000)
#define NEXT_CONN_PARAMS_UPDATE_DELAY APP_TIMER_TICKS(30000)
#define MAX_CONN_PARAMS_UPDATE_COUNT 3

I understand these conform to Apple's requirements. We do NOT need high throughput and send only small amounts of data (<20 bytes) between central and peripherals. In sdk_config.h, NRF_SDH_BLE_GATT_MAX_MTU_SIZE is 23. In ble_evt_handler(), on BLE_GAP_EVT_CONNECTED, I set transmit power level to 4db:

err_code = sd_ble_gap_tx_power_set(BLE_GAP_TX_POWER_ROLE_CONN, m_conn_handle, 4);

Using a BLE sniffer, for iOS 13.5.1 on an iPhone SE (2020 model) we see the following CONNECT_REQ packet:

Bluetooth Low Energy Link Layer
    Access Address: 0x8e89bed6
    Packet Header: 0x22c5 (PDU Type: CONNECT_REQ, ChSel: #1, TxAdd: Random, RxAdd: Random)
    Initator Address: 5c:d1:b4:78:4e:43 (5c:d1:b4:78:4e:43)
    Advertising Address: d5:88:07:32:7a:ad (d5:88:07:32:7a:ad)
    Link Layer Data
        Access Address: 0x50654a99
        CRC Init: 0x28ce8e
        Window Size: 3 (3.75 msec)
        Window Offset: 7 (8.75 msec)
        Interval: 24 (30 msec)
        Latency: 0
        Timeout: 72 (720 msec)
        Channel Map: ff07c0ff1f
            .... ...1 = RF Channel 1 (2404 MHz - Data - 0): True
            .... ..1. = RF Channel 2 (2406 MHz - Data - 1): True
            .... .1.. = RF Channel 3 (2408 MHz - Data - 2): True
            .... 1... = RF Channel 4 (2410 MHz - Data - 3): True
            ...1 .... = RF Channel 5 (2412 MHz - Data - 4): True
            ..1. .... = RF Channel 6 (2414 MHz - Data - 5): True
            .1.. .... = RF Channel 7 (2416 MHz - Data - 6): True
            1... .... = RF Channel 8 (2418 MHz - Data - 7): True
            .... ...1 = RF Channel 9 (2420 MHz - Data - 8): True
            .... ..1. = RF Channel 10 (2422 MHz - Data - 9): True
            .... .1.. = RF Channel 11 (2424 MHz - Data - 10): True
            .... 0... = RF Channel 13 (2428 MHz - Data - 11): False
            ...0 .... = RF Channel 14 (2430 MHz - Data - 12): False
            ..0. .... = RF Channel 15 (2432 MHz - Data - 13): False
            .0.. .... = RF Channel 16 (2434 MHz - Data - 14): False
            0... .... = RF Channel 17 (2436 MHz - Data - 15): False
            .... ...0 = RF Channel 18 (2438 MHz - Data - 16): False
            .... ..0. = RF Channel 19 (2440 MHz - Data - 17): False
            .... .0.. = RF Channel 20 (2442 MHz - Data - 18): False
            .... 0... = RF Channel 21 (2444 MHz - Data - 19): False
            ...0 .... = RF Channel 22 (2446 MHz - Data - 20): False
            ..0. .... = RF Channel 23 (2448 MHz - Data - 21): False
            .1.. .... = RF Channel 24 (2450 MHz - Data - 22): True
            1... .... = RF Channel 25 (2452 MHz - Data - 23): True
            .... ...1 = RF Channel 26 (2454 MHz - Data - 24): True
            .... ..1. = RF Channel 27 (2456 MHz - Data - 25): True
            .... .1.. = RF Channel 28 (2458 MHz - Data - 26): True
            .... 1... = RF Channel 29 (2460 MHz - Data - 27): True
            ...1 .... = RF Channel 30 (2462 MHz - Data - 28): True
            ..1. .... = RF Channel 31 (2464 MHz - Data - 29): True
            .1.. .... = RF Channel 32 (2466 MHz - Data - 30): True
            1... .... = RF Channel 33 (2468 MHz - Data - 31): True
            .... ...1 = RF Channel 34 (2470 MHz - Data - 32): True
            .... ..1. = RF Channel 35 (2472 MHz - Data - 33): True
            .... .1.. = RF Channel 36 (2474 MHz - Data - 34): True
            .... 1... = RF Channel 37 (2476 MHz - Data - 35): True
            ...1 .... = RF Channel 38 (2478 MHz - Data - 36): True
            ..0. .... = RF Channel 0 (2402 MHz - Reserved for Advertising - 37): False
            .0.. .... = RF Channel 12 (2426 MHz - Reserved for Advertising - 38): False
            0... .... = RF Channel 39 (2480 MHz - Reserved for Advertising - 39): False
        ...0 1111 = Hop: 15
        001. .... = Sleep Clock Accuracy: 151 ppm to 250 ppm (1)
    CRC: 0x419071

So we have 9 peripherals, 3 connected to one iOS device, 3 connected to another iOS device, and 3 connected to yet another iOS device. I'm noticing random disconnects. On the iOS side, centralManager(_:didDisconnectPeripheral:error:) is reporting error 6 which is "The connection has timed out.".

Have we bumped into some of the practical limits of BLE? I did a sniffer trace that captured a random disconnect. See attached (Wireshark with nRF52 DK and nRF Sniffer 3.0.0). I notice a lot of LL_CHANNEL_MAP_REQ packets, but I don't have much knowledge of this level of BLE. Is there anything we can do to increase connection stability? We request a 4 second supervisor timeout but the central chooses 720 milliseconds. Use higher min and max connection interval? Our central app generally uses writeWithRespnse when writing characteristic values.

Appreciate any information, guidance.

Many thanks,

Tim

5238.sniffer trace.pcapng

Parents
  • Hello Tim,

    I have looked at the sniffer log that you provided, and it looks like the nRF (the peripheral/slave) is not responding, and hence the disconnect, as you probably know.

    What does the nRF behave like in this case? Have you tried to debug the application there? Do you get the disconnected event on the nRF as well, or does it for some reason reset? If you receive the disconnect event, what reason is it pointing to?

    You can find the disconnect reason in the BLE_GAP_EVT_DISCONNECTED event by using:

    NRF_LOG_INFO("Disconnected, reason: 0x%02x", p_ble_evt->evt.gap_evt.params.disconnected.reason);

    What does it say?

    Best regards,

    Edvin

  • Thanks Edvin. It will take some time to set up a debug environment in the field, but I can do that and find out if the BLE_GAP_EVT_DISCONNECTED event occurs on the peripheral and what the disconnect reason is.

    In the meantime, I would be glad for any general guidance on creating a most stable link connection. Given that data throughput is not important, and that only small amounts of data are written (in either direction), what would be suggested connection parameters? Any other helpful considerations?

    In reading the Timeslot API documentation, I understand that using it can affect performance of the SoftDevice, so I wonder if it is a factor in what I am experiencing.

    I've attached another sniffer trace. It appears the disconnect occurs around time 103.733 as the peripheral/slave begins advertising again at 104.661. It seems that throughout duration of the connection, occasional Empty PDU packets from the slave are not seen by the sniffer. For example, at 15.408, 15.497, 15.588, etc. Prior to disconnect at 103.733, dropped slave Empty PDU packets become more frequent. At 98.002 there are 3 missing slave Empty PDU packets. At 100.852 there are 8 dropped packets, then more and more dropped packets until disconnect around 103.733. Would this suggest an interference issue? Maybe as clocks of other nearby peripherals drift, and at some point the interference becomes intolerable? Again, I have very little experience here, so maybe unhelpful speculation.

    Many thanks,

    Tim

    sniffer trace-2.pcapng

  • Tim said:
    Given throughput is not important, power consumption is not important (peripheral devices powered by 2 AA batteries), but reliable use with Timeslot API is important, what min/max connection intervals would you recommend?

     If your central is always an iPhone it doesn't really matter. Set min = 30ms and max = 100ms. You will get 30ms either way. If you set max=100ms, and Apple one day in the future supports >30ms, you will save some battery.

     

    Tim said:
    How do I attain a higher connection interval if  iOS always denies and uses 30 ms?

    You can't. It is always the Central that controls this parameter. As you mentioned in the previous reply, you can request changes, but it is the central that has the final saying. The only thing the peripheral can do is to disconnect if you don't get the connection parameters you desire, but I guess you don't really want to do that. 

     

    Tim said:
    Is there a way a peripheral application can read the connection interval in use for debug logging?

     I guess it says in the connected event and in the connection parameter update event. 

    Check out the parameters in this struckt in the BLE_GAP_EVT_CONNECTED:

    p_ble_evt->evt.gap_evt.params.connected.conn_params... [min_conn_interval, max_conn_interval_slave_latency, conn_sup_timeout]

    And check out the parameters in this struct in BLE_GAP_EVT_CONN_PARAM_UPDATE event:

    p_ble_evt->evt.gap_evt.params.conn_param_update.conn_params... [the same parameters as the BLE_GAP_EVT_CONNECTED event]

     

    Tim said:
    but then the LL_PHY_UPDATE_IND packets from the master show False for all modes.

     I believe this packet is just an "abort" message for the phy change. It also says "instance: 0". Instance refers to the event counter that these changes shall be enforced. If you want to monitor the changes from the application, check out the BLE_GAP_EVT_PHY_UPDATE_REQUEST and BLE_GAP_EVT_PHY_UPDATE_REQUEST events (look for the parameters like you do for the connected and conn_param_update events).

    The BLE_GAP_EVT_PHY_UPDATE_REQUEST event will contain the phy parameters that the central requests, and if you get the BLE_GAP_EVT_PHY_UPDATE event, it means that the PHY is updated, and the event contains the new PHY.

    Best regards,

    Edvin

  • Hi Edvin,

    I've discovered that with iPhone/iOS, if slave requests minimum connection interval is 30ms, then you'll always get 30ms regardless of what maximum interval is. However, if minimum interval is 45ms or higher (needs to be multiple of 15 to meet Apple requirements), then you'll get whatever the specified maximum interval is. I've tested on both old iPhone (4s) and new (iPhone SE (2020)) and observe this behaviour.

    I've attached two traces. conn_intervals_30-105.pcapng uses min 30, max 105. Notice no connection parameter update request, and the initial CONNECT_REQ packet shows 30ms interval.

    conn_intervals_30-105.pcapng

    conn_intervals_45-105.pcapng

    conn_intervals_45-105.pcapng uses min 45, max 105. Notice connection parameter request from the slave at packet no. 734, then the response at no. 735 with move result "Accepted". Then no. 757 LL_CONNECTION_UPDATE_REQ showing interval of 84, which is 105ms.

    In most recent tests the connection instability previously experienced was gone. Two things changed:

    - used connection interval of 120ms

    - switched from 2MBPS to 1MBPS PHY

    Given high throughput is not a requirement, thought to go with min 60 and max 105. Reasonable?

    What is the reason for the 5 second delay before calling sd_ble_gap_conn_param_update() in ble_app_uart example of SDK 15.3.0 (FIRST_CONN_PARAMS_UPDATE_DELAY = 5s)?

    Many thanks,

    Tim

  • Hello Tim,

    That is good. I thought iPhones just set it to 30ms no matter what connection interval you used. So if you are able to set it higher, it will reduce the current consumption. Please note that it will also increase the latency of the link. 

    Remember that when you increase the connection interval, you may also want to increase the supervision timeout. If you have a connection interval of 60ms, and a supervision timeout of 600ms (just random numbers) you can miss 10 packets before a disconnect, while if you increase the connection interval to 120ms, you can only miss 5. 

     

    Tim said:
    What is the reason for the 5 second delay before calling sd_ble_gap_conn_param_update() in ble_app_uart example of SDK 15.3.0 (FIRST_CONN_PARAMS_UPDATE_DELAY = 5s)?

     I don't remember exactly the reason, but there is something that we want to let finish before we start this. I believe it is the service discovery that we want the central to be able to perform before we change the connection interval.

    Best regards,

    Edvin

  • Thanks Edvin. Testing has been going well. Thanks to your help, I have a better understanding of some of the interaction between central and peripheral. Some of my current assumptions:

    • A link between central and peripheral is always initially PHY 1MBPS. Then if the central supports it (i.e. BLE 5.0 or later), it will send a BLE_GAP_EVT_PHY_UPDATE_REQUEST to the peripheral and the peripheral can update PHY to any of the available options (AUTO, 1MBPS, 1MBPS, CODED). For our application, range is more important than throughput, so we set to 1MBPS.
    • It seems that the initial connection interval when connecting peripheral with iOS central is always 30 ms. If connection parameter negotiation offers minimum interval of 30 ms with any maximum (always needs to be multiple of 15 ms), iOS will always select 30 ms. However, if peripheral offers minimum greater than 30 ms (e.g. 45, 60, 75, etc.) and a maximum greater than that, iOS will always choose the maximum. I've tested up to a maximum of 120 ms.
    • A connection event is always 7.5 ms. So a larger connection interval means more idle time between connection events. This can be helpful when BLE coexists with timeslot API activity, as in our case.
    • Yes, this post says connection parameter negotiation occurs 5 seconds after connection (in many examples) to be sure service discovery has completed.

    This post says S110 has 6 tx buffers. Is this true also for S112 v6.1.1? Rarely does our application call sd_ble_gatts_hvx() more than twice in quick succession.

    Recently we've been testing with connection interval of 120 ms. Our application uses the timeslot API for radio use between BLE activity. I presume that a larger connection interval makes it easier for time slots to be successfully scheduled, correct?

    Many thanks Edvin.

    Tim

  • Hello,

    Most of this seems right, but a few corrections:

     

    Tim said:
    A connection event is always 7.5 ms. So a larger connection interval means more idle time between connection events. This can be helpful when BLE coexists with timeslot API activity, as in our case.

     This is not necessarily true. Typically, a connection event is only a couple of ms (with little payload), and up to the entire connection event, depending on the connection parameters, and the payload.

    You can check it out here:

    https://devzone.nordicsemi.com/nordic/power/w/opp/2/online-power-profiler-for-ble

     

    Tim said:
    S110 has 6 tx buffers. Is this true also for S112 v6.1.1?

     It really depends on the packet size. You can fill it with one long notification as well.

    While I believe it is possible to increase this queue size, that is rarely the solution, as it will just move the queue from one location to another. Keep track of what packets you have been able to queue. As long as the function you use to queue returns NRF_SUCCESS, it is queued and will be sent. You can forget the payload from your application side. If it returns NRF_ERROR_RESOURCES, it means that the packet didn't fit in the queue, and it was not queued. In this case, you must wait for the next BLE_GATTS_EVT_HVN_TX_COMPLETE event. This means that a packet from your softdevice queue was acked, and it has room for more data. At this point you can call the function that calls sd_ble_gatts_hvx() function again.

    Best regards,

    Edvin

Reply
  • Hello,

    Most of this seems right, but a few corrections:

     

    Tim said:
    A connection event is always 7.5 ms. So a larger connection interval means more idle time between connection events. This can be helpful when BLE coexists with timeslot API activity, as in our case.

     This is not necessarily true. Typically, a connection event is only a couple of ms (with little payload), and up to the entire connection event, depending on the connection parameters, and the payload.

    You can check it out here:

    https://devzone.nordicsemi.com/nordic/power/w/opp/2/online-power-profiler-for-ble

     

    Tim said:
    S110 has 6 tx buffers. Is this true also for S112 v6.1.1?

     It really depends on the packet size. You can fill it with one long notification as well.

    While I believe it is possible to increase this queue size, that is rarely the solution, as it will just move the queue from one location to another. Keep track of what packets you have been able to queue. As long as the function you use to queue returns NRF_SUCCESS, it is queued and will be sent. You can forget the payload from your application side. If it returns NRF_ERROR_RESOURCES, it means that the packet didn't fit in the queue, and it was not queued. In this case, you must wait for the next BLE_GATTS_EVT_HVN_TX_COMPLETE event. This means that a packet from your softdevice queue was acked, and it has room for more data. At this point you can call the function that calls sd_ble_gatts_hvx() function again.

    Best regards,

    Edvin

Children
No Data
Related