aboutsummaryrefslogtreecommitdiffstats
path: root/host
diff options
context:
space:
mode:
Diffstat (limited to 'host')
-rw-r--r--host/include/uhd/transport/nirio/nirio_fifo.h136
-rw-r--r--host/include/uhd/transport/nirio/nirio_fifo.ipp189
-rw-r--r--host/lib/transport/nirio/niriok_proxy_impl_v2.cpp14
-rw-r--r--host/lib/transport/nirio_zero_copy.cpp2
4 files changed, 295 insertions, 46 deletions
diff --git a/host/include/uhd/transport/nirio/nirio_fifo.h b/host/include/uhd/transport/nirio/nirio_fifo.h
index a0cd7149f..f6eb97ce2 100644
--- a/host/include/uhd/transport/nirio/nirio_fifo.h
+++ b/host/include/uhd/transport/nirio/nirio_fifo.h
@@ -28,6 +28,7 @@
#include <string>
#include <boost/thread/recursive_mutex.hpp>
#include <boost/thread/thread.hpp>
+#include <boost/atomic/atomic.hpp>
namespace uhd { namespace niusrprio {
@@ -43,29 +44,91 @@ class nirio_fifo : private boost::noncopyable
public:
typedef boost::shared_ptr< nirio_fifo<data_t> > sptr;
+ typedef enum {
+ MINIMIZE_LATENCY,
+ MAXIMIZE_THROUGHPUT
+ } fifo_optimization_option_t;
+
+ /*!
+ * Initializing constructor
+ * \param riok_proxy Proxy to the kernel driver
+ * \param direction Direction of the fifo (INPUT_FIFO [from device] or OUTPUT_FIFO [to device])
+ * \param name Name of the fifo
+ * \param fifo_instance Instance of the fifo as defined by RIO FPGA target
+ */
nirio_fifo(
niriok_proxy::sptr riok_proxy,
- fifo_direction_t direction,
+ const fifo_direction_t direction,
const std::string& name,
- uint32_t fifo_instance);
+ const uint32_t fifo_instance);
virtual ~nirio_fifo();
+ /*!
+ * Configures the characterists of this DMA FIFO, allocates DMA buffer and maps it to user mode
+ * \param requested_depth Desired size of DMA FIFO buffer on host (in elements)
+ * \param frame_size_in_elements Size of a single frame (smallest transfer block) for this FIFO
+ * \param actual_depth Receives the size (in elements) of the allocated DMA buffer
+ * \param actual_size Receives the size (in bytes) of the allocated DMA buffer
+ * \param fifo_optimization_option FIFO acquire policy (MINIMIZE_LATENCY [fetch what was there last time] or MAXIMIZE_THROUGHPUT [fetch all available])
+ * \return status
+ */
nirio_status initialize(
const size_t requested_depth,
+ const size_t frame_size_in_elements,
size_t& actual_depth,
- size_t& actual_size);
+ size_t& actual_size,
+ const fifo_optimization_option_t fifo_optimization_option = MINIMIZE_LATENCY);
+ /*!
+ * Stops FIFO if started, releases any acquired blocks, and unmaps the DMA buffer
+ */
void finalize();
+ /*!
+ * Name accessor
+ * \return FIFO name
+ */
inline const std::string& get_name() const { return _name; }
+
+ /*!
+ * Channel/Instance accessor
+ * \return FIFO channel (instance)
+ */
inline uint32_t get_channel() const { return _fifo_channel; }
+
+ /*!
+ * Direction accessor
+ * \return FIFO direction
+ */
inline fifo_direction_t get_direction() const { return _fifo_direction; }
+
+ /*!
+ * Type accessor
+ * \return FIFO element type
+ */
inline nirio_scalar_type_t get_scalar_type() const { return _datatype_info.scalar_type; }
+ /*!
+ * Starts the DMA transfer between host and device, pre-acquires any available blocks
+ * \return status
+ */
nirio_status start();
+ /*!
+ * Stops the DMA transfer between host and device, releases any acquired blocks
+ * \return status
+ */
nirio_status stop();
+ /*!
+ * Acquires space in the DMA buffer so it can be written by the host (output) or read by the host (input)
+ * \param elements Receives the address of the acquired block (pointer to mapped zero-copy DMA buffer)
+ * \param elements_requested Size (in elements) of the block to acquire
+ * \param timeout The amount of time (in ms) to wait for the elements to become available in the DMA buffer
+ * \param elements_acquired Receives the number of DMA buffer elements actually acquired
+ * \param elements_remaining Receives the number of DMA buffer elements available to be acquired by the host
+ * \return status
+ */
nirio_status acquire(
data_t*& elements,
const size_t elements_requested,
@@ -73,26 +136,80 @@ public:
size_t& elements_acquired,
size_t& elements_remaining);
+ /*!
+ * Releases space in the DMA buffer so it can be read by the device (output) or written by the device (input)
+ * \param elements Size (in elements) of the block to release.
+ * \return status
+ */
nirio_status release(const size_t elements);
+ /*!
+ * Reads data from the DMA FIFO into the provided buffer
+ * \param buf The buffer into which to read data from the DMA FIFO
+ * \param num_elements Size (in elements) of the data to read
+ * \param timeout The amount of time (in ms) to wait for the elements to become available in the DMA buffer
+ * \param num_read Receives the number of DMA buffer elements actually read
+ * \param num_remaining Receives the number of DMA buffer elements available be read by the host
+ * \return status
+ */
nirio_status read(
data_t* buf,
const uint32_t num_elements,
- uint32_t timeout,
+ const uint32_t timeout,
uint32_t& num_read,
uint32_t& num_remaining);
+ /*!
+ * Writes data from the DMA FIFO
+ * \param buf The buffer containing data to be written to the DMA FIFO
+ * \param num_elements Size (in elements) of the data to write
+ * \param timeout The amount of time (in ms) to wait for the elements to become available in the DMA buffer
+ * \param num_remaining Receives the number of DMA buffer elements available be written by the host
+ * \return status
+ */
nirio_status write(
const data_t* buf,
const uint32_t num_elements,
- uint32_t timeout,
+ const uint32_t timeout,
uint32_t& num_remaining);
private: //Methods
+
+ /*!
+ * datatype info accessor
+ * \return datatype info
+ */
datatype_info_t _get_datatype_info();
+
+ /*!
+ * Queries the total transfer count so far between host and device
+ * \param transfer_count Receives the value from total transfer count register
+ * \return status
+ */
nirio_status _get_transfer_count(uint64_t& transfer_count);
+
+ /*!
+ * Sleeps until expected transfer time has elapsed and checks total transfer count to see if transfer completed
+ * \param timeout_ms The amount of time (in ms) to wait for the elements to become available in the DMA buffer
+ * \return status
+ */
nirio_status _ensure_transfer_completed(uint32_t timeout_ms);
+
+ /*!
+ * Conducts the low level operations to reserve DMA buffer space from RIO kernel driver
+ * \param elements_requested The minimum number of elements to acquire
+ * \param timeout_in_ms The amount of time (in ms) to wait for the elements to become available in the DMA buffer
+ * \param fifo_optimization_option FIFO acquire policy (MINIMIZE_LATENCY [fetch what was there last time] or MAXIMIZE_THROUGHPUT [fetch all available])
+ * \param status status chaining variable
+ * \return Whether acquisition of requested elements (or more) was successful.
+ */
+ bool _acquire_block_from_rio_buffer(
+ size_t elements_requested,
+ uint64_t timeout_in_ms,
+ const fifo_optimization_option_t fifo_optimization_option,
+ nirio_status& status);
+
private: //Members
enum fifo_state_t {
UNMAPPED, MAPPED, STARTED
@@ -103,7 +220,8 @@ private: //Members
uint32_t _fifo_channel;
datatype_info_t _datatype_info;
fifo_state_t _state;
- size_t _acquired_pending;
+ size_t _remaining_in_claimed_block;
+ size_t _remaining_acquirable_elements;
nirio_driver_iface::rio_mmap_t _mem_map;
boost::recursive_mutex _mutex;
niriok_proxy::sptr _riok_proxy_ptr;
@@ -111,6 +229,12 @@ private: //Members
uint64_t _expected_xfer_count;
uint32_t _dma_base_addr;
+ data_t* _elements_buffer;
+ size_t _actual_depth_in_elements;
+ boost::atomic<size_t> _total_elements_acquired;
+ size_t _frame_size_in_elements;
+ fifo_optimization_option_t _fifo_optimization_option;
+
static const uint32_t FIFO_LOCK_TIMEOUT_IN_MS = 5000;
};
diff --git a/host/include/uhd/transport/nirio/nirio_fifo.ipp b/host/include/uhd/transport/nirio/nirio_fifo.ipp
index a7e2bb2c5..5669f259f 100644
--- a/host/include/uhd/transport/nirio/nirio_fifo.ipp
+++ b/host/include/uhd/transport/nirio/nirio_fifo.ipp
@@ -26,19 +26,25 @@
template <typename data_t>
nirio_fifo<data_t>::nirio_fifo(
niriok_proxy::sptr riok_proxy,
- fifo_direction_t direction,
+ const fifo_direction_t direction,
const std::string& name,
- uint32_t fifo_instance) :
+ const uint32_t fifo_instance) :
_name(name),
_fifo_direction(direction),
_fifo_channel(fifo_instance),
_datatype_info(_get_datatype_info()),
_state(UNMAPPED),
- _acquired_pending(0),
+ _remaining_in_claimed_block(0),
+ _remaining_acquirable_elements(0),
_mem_map(),
_riok_proxy_ptr(riok_proxy),
_expected_xfer_count(0),
- _dma_base_addr(0)
+ _dma_base_addr(0),
+ _elements_buffer(NULL),
+ _actual_depth_in_elements(0),
+ _total_elements_acquired(0),
+ _frame_size_in_elements(0),
+ _fifo_optimization_option(MINIMIZE_LATENCY)
{
nirio_status status = 0;
nirio_status_chain(_riok_proxy_ptr->set_attribute(RIO_ADDRESS_SPACE, BUS_INTERFACE), status);
@@ -58,8 +64,10 @@ nirio_fifo<data_t>::~nirio_fifo()
template <typename data_t>
nirio_status nirio_fifo<data_t>::initialize(
const size_t requested_depth,
+ const size_t frame_size_in_elements,
size_t& actual_depth,
- size_t& actual_size)
+ size_t& actual_size,
+ const fifo_optimization_option_t fifo_optimization_option)
{
nirio_status status = NiRio_Status_Success;
if (!_riok_proxy_ptr) return NiRio_Status_ResourceNotInitialized;
@@ -67,6 +75,9 @@ nirio_status nirio_fifo<data_t>::initialize(
if (_state == UNMAPPED) {
+ _frame_size_in_elements = frame_size_in_elements;
+ _fifo_optimization_option = fifo_optimization_option;
+
uint32_t actual_depth_u32 = 0;
uint32_t actual_size_u32 = 0;
@@ -83,6 +94,7 @@ nirio_status nirio_fifo<data_t>::initialize(
if (nirio_status_fatal(status)) return status;
actual_depth = static_cast<size_t>(actual_depth_u32);
+ _actual_depth_in_elements = actual_depth;
actual_size = static_cast<size_t>(actual_size_u32);
status = _riok_proxy_ptr->map_fifo_memory(_fifo_channel, actual_size, _mem_map);
@@ -110,6 +122,89 @@ void nirio_fifo<data_t>::finalize()
}
}
+
+template <typename data_t>
+bool nirio_fifo<data_t>::_acquire_block_from_rio_buffer(
+ size_t elements_requested,
+ uint64_t timeout_in_ms,
+ const fifo_optimization_option_t fifo_optimization_option,
+ nirio_status& status)
+{
+ uint32_t elements_acquired_u32 = 0;
+ uint32_t elements_remaining_u32 = 0;
+ size_t elements_to_request = 0;
+ void* elements_buffer = NULL;
+ char context_buffer[64];
+
+ if (fifo_optimization_option == MAXIMIZE_THROUGHPUT)
+ {
+ // We'll maximize throughput by acquiring all the data that is available
+ // But this comes at the cost of an extra wait_on_fifo in which we query
+ // the total available by requesting 0.
+
+ // first, see how many are available to acquire
+ // by trying to acquire 0
+ nirio_status_chain(_riok_proxy_ptr->wait_on_fifo(
+ _fifo_channel,
+ 0, // elements requested
+ static_cast<uint32_t>(_datatype_info.scalar_type),
+ _datatype_info.width * 8,
+ 0, // timeout
+ _fifo_direction == OUTPUT_FIFO,
+ elements_buffer,
+ elements_acquired_u32,
+ elements_remaining_u32),
+ status);
+
+ // acquire the maximum possible elements- all remaining
+ // yet limit to a multiple of the frame size
+ // (don't want to acquire partial frames)
+ elements_to_request = elements_remaining_u32 - (elements_remaining_u32 % _frame_size_in_elements);
+
+ // the next call to wait_on_fifo can have a 0 timeout since we
+ // know there is at least as much data as we will request available
+ timeout_in_ms = 0;
+ }
+ else
+ {
+ // fifo_optimization_option == MINIMIZE_LATENCY
+ // acquire either the minimum amount (frame size) or the amount remaining from the last call
+ // (coerced to a multiple of frames)
+ elements_to_request = std::max(
+ elements_requested,
+ (_remaining_acquirable_elements - (_remaining_acquirable_elements % _frame_size_in_elements)));
+ }
+
+ nirio_status_chain(_riok_proxy_ptr->wait_on_fifo(
+ _fifo_channel,
+ elements_to_request,
+ static_cast<uint32_t>(_datatype_info.scalar_type),
+ _datatype_info.width * 8,
+ timeout_in_ms,
+ _fifo_direction == OUTPUT_FIFO,
+ elements_buffer,
+ elements_acquired_u32,
+ elements_remaining_u32),
+ status);
+
+ if (nirio_status_not_fatal(status))
+ {
+ _remaining_acquirable_elements = static_cast<size_t>(elements_remaining_u32);
+
+ if (elements_acquired_u32 > 0)
+ {
+ _total_elements_acquired += static_cast<size_t>(elements_acquired_u32);
+ _remaining_in_claimed_block = static_cast<size_t>(elements_acquired_u32);
+ _elements_buffer = static_cast<data_t*>(elements_buffer);
+ }
+
+ return true;
+ }
+
+ return false;
+}
+
+
template <typename data_t>
nirio_status nirio_fifo<data_t>::start()
{
@@ -122,11 +217,24 @@ nirio_status nirio_fifo<data_t>::start()
//Do nothing. Already started.
} else if (_state == MAPPED) {
+ _total_elements_acquired = 0;
+ _remaining_in_claimed_block = 0;
+ _remaining_acquirable_elements = 0;
+
status = _riok_proxy_ptr->start_fifo(_fifo_channel);
+
if (nirio_status_not_fatal(status)) {
_state = STARTED;
- _acquired_pending = 0;
_expected_xfer_count = 0;
+
+ if (_fifo_direction == OUTPUT_FIFO)
+ {
+ // pre-acquire a block of data
+ // (should be entire DMA buffer at this point)
+
+ // requesting 0 elements, but it will acquire all since MAXIMUM_THROUGHPUT
+ _acquire_block_from_rio_buffer(0, 1000, MAXIMIZE_THROUGHPUT, status);
+ }
}
} else {
status = NiRio_Status_ResourceNotInitialized;
@@ -143,7 +251,12 @@ nirio_status nirio_fifo<data_t>::stop()
boost::unique_lock<boost::recursive_mutex> lock(_mutex);
if (_state == STARTED) {
- if (_acquired_pending > 0) release(_acquired_pending);
+
+ // release any remaining acquired elements
+ if (_total_elements_acquired > 0) release(_total_elements_acquired);
+ _total_elements_acquired = 0;
+ _remaining_in_claimed_block = 0;
+ _remaining_acquirable_elements = 0;
status = _riok_proxy_ptr->stop_fifo(_fifo_channel);
@@ -167,34 +280,44 @@ nirio_status nirio_fifo<data_t>::acquire(
boost::unique_lock<boost::recursive_mutex> lock(_mutex);
if (_state == STARTED) {
- uint32_t elements_acquired_u32 = 0;
- uint32_t elements_remaining_u32 = 0;
- void* elements_buffer = static_cast<void*>(elements);
- status = _riok_proxy_ptr->wait_on_fifo(
- _fifo_channel,
- static_cast<uint32_t>(elements_requested),
- static_cast<uint32_t>(_datatype_info.scalar_type),
- _datatype_info.width * 8,
- timeout,
- _fifo_direction == OUTPUT_FIFO,
- elements_buffer,
- elements_acquired_u32,
- elements_remaining_u32);
- if (nirio_status_not_fatal(status)) {
- elements = static_cast<data_t*>(elements_buffer);
- elements_acquired = static_cast<size_t>(elements_acquired_u32);
- elements_remaining = static_cast<size_t>(elements_remaining_u32);
- _acquired_pending = elements_acquired;
-
- if (UHD_NIRIO_RX_FIFO_XFER_CHECK_EN &&
- _riok_proxy_ptr->get_rio_quirks().rx_fifo_xfer_check_en() &&
- get_direction() == INPUT_FIFO
- ) {
+ if (_remaining_in_claimed_block == 0)
+ {
+
+ // so acquire some now
+ if (!_acquire_block_from_rio_buffer(
+ elements_requested,
+ timeout,
+ _fifo_optimization_option,
+ status))
+ {
+ elements_acquired = 0;
+ elements_remaining = _remaining_acquirable_elements;
+ return status;
+ }
+
+
+ if (get_direction() == INPUT_FIFO &&
+ UHD_NIRIO_RX_FIFO_XFER_CHECK_EN &&
+ _riok_proxy_ptr->get_rio_quirks().rx_fifo_xfer_check_en())
+ {
_expected_xfer_count += static_cast<uint64_t>(elements_requested * sizeof(data_t));
status = _ensure_transfer_completed(timeout);
}
}
+
+ if (nirio_status_not_fatal(status))
+ {
+ // Assign the request the proper area of the DMA FIFO buffer
+ elements = _elements_buffer;
+ elements_acquired = std::min(_remaining_in_claimed_block, elements_requested);
+ _remaining_in_claimed_block -= elements_acquired;
+ elements_remaining = _remaining_in_claimed_block + _remaining_acquirable_elements;
+
+ // Advance the write pointer forward in the acquired elements buffer
+ _elements_buffer += elements_acquired;
+ }
+
} else {
status = NiRio_Status_ResourceNotInitialized;
}
@@ -214,7 +337,7 @@ nirio_status nirio_fifo<data_t>::release(const size_t elements)
status = _riok_proxy_ptr->grant_fifo(
_fifo_channel,
static_cast<uint32_t>(elements));
- _acquired_pending = 0;
+ _total_elements_acquired -= elements;
} else {
status = NiRio_Status_ResourceNotInitialized;
}
@@ -226,7 +349,7 @@ template <typename data_t>
nirio_status nirio_fifo<data_t>::read(
data_t* buf,
const uint32_t num_elements,
- uint32_t timeout,
+ const uint32_t timeout,
uint32_t& num_read,
uint32_t& num_remaining)
{
@@ -257,7 +380,7 @@ template <typename data_t>
nirio_status nirio_fifo<data_t>::write(
const data_t* buf,
const uint32_t num_elements,
- uint32_t timeout,
+ const uint32_t timeout,
uint32_t& num_remaining)
{
nirio_status status = NiRio_Status_Success;
diff --git a/host/lib/transport/nirio/niriok_proxy_impl_v2.cpp b/host/lib/transport/nirio/niriok_proxy_impl_v2.cpp
index f75de01e1..4e88b8a6e 100644
--- a/host/lib/transport/nirio/niriok_proxy_impl_v2.cpp
+++ b/host/lib/transport/nirio/niriok_proxy_impl_v2.cpp
@@ -759,14 +759,14 @@ namespace uhd { namespace niusrprio
in.timeout = timeout;
in.status = NiRio_Status_Success;
- ioctl_status =
+ ioctl_status =
nirio_driver_iface::rio_ioctl(
- _device_handle,
- IOCTL_TRANSPORT_FIFO_WAIT,
- &in,
- sizeof(in),
- &out,
- sizeof(out));
+ _device_handle,
+ IOCTL_TRANSPORT_FIFO_WAIT,
+ &in,
+ sizeof(in),
+ &out,
+ sizeof(out));
if (nirio_status_fatal(ioctl_status)) return ioctl_status;
data_pointer = reinterpret_cast<void*>(out.elements);
diff --git a/host/lib/transport/nirio_zero_copy.cpp b/host/lib/transport/nirio_zero_copy.cpp
index e7a37eb08..81ba2a20e 100644
--- a/host/lib/transport/nirio_zero_copy.cpp
+++ b/host/lib/transport/nirio_zero_copy.cpp
@@ -182,11 +182,13 @@ public:
nirio_status_chain(
_recv_fifo->initialize(
(_xport_params.recv_frame_size*_xport_params.num_recv_frames)/sizeof(fifo_data_t),
+ _xport_params.recv_frame_size / sizeof(fifo_data_t),
actual_depth, actual_size),
status);
nirio_status_chain(
_send_fifo->initialize(
(_xport_params.send_frame_size*_xport_params.num_send_frames)/sizeof(fifo_data_t),
+ _xport_params.send_frame_size / sizeof(fifo_data_t),
actual_depth, actual_size),
status);