@@ -58,11 +58,11 @@ struct uart_sam0_dev_data {
58
58
uart_callback_t async_cb ;
59
59
void * async_cb_data ;
60
60
61
- struct k_delayed_work tx_timeout_work ;
61
+ struct k_work_delayable tx_timeout_work ;
62
62
const uint8_t * tx_buf ;
63
63
size_t tx_len ;
64
64
65
- struct k_delayed_work rx_timeout_work ;
65
+ struct k_work_delayable rx_timeout_work ;
66
66
size_t rx_timeout_time ;
67
67
size_t rx_timeout_chunk ;
68
68
uint32_t rx_timeout_start ;
@@ -130,7 +130,7 @@ static void uart_sam0_dma_tx_done(const struct device *dma_dev, void *arg,
130
130
(struct uart_sam0_dev_data * const ) arg ;
131
131
const struct device * dev = dev_data -> dev ;
132
132
133
- k_delayed_work_cancel (& dev_data -> tx_timeout_work );
133
+ k_work_cancel_delayable (& dev_data -> tx_timeout_work );
134
134
135
135
int key = irq_lock ();
136
136
@@ -361,7 +361,7 @@ static void uart_sam0_rx_timeout(struct k_work *work)
361
361
*/
362
362
if (dev_data -> rx_timeout_from_isr ) {
363
363
dev_data -> rx_timeout_from_isr = false;
364
- k_delayed_work_submit (& dev_data -> rx_timeout_work ,
364
+ k_work_reschedule (& dev_data -> rx_timeout_work ,
365
365
K_MSEC (dev_data -> rx_timeout_chunk ));
366
366
irq_unlock (key );
367
367
return ;
@@ -383,7 +383,7 @@ static void uart_sam0_rx_timeout(struct k_work *work)
383
383
uint32_t remaining = MIN (dev_data -> rx_timeout_time - elapsed ,
384
384
dev_data -> rx_timeout_chunk );
385
385
386
- k_delayed_work_submit (& dev_data -> rx_timeout_work ,
386
+ k_work_reschedule (& dev_data -> rx_timeout_work ,
387
387
K_MSEC (remaining ));
388
388
}
389
389
@@ -570,8 +570,8 @@ static int uart_sam0_init(const struct device *dev)
570
570
return - ENODEV ;
571
571
}
572
572
573
- k_delayed_work_init (& dev_data -> tx_timeout_work , uart_sam0_tx_timeout );
574
- k_delayed_work_init (& dev_data -> rx_timeout_work , uart_sam0_rx_timeout );
573
+ k_work_init_delayable (& dev_data -> tx_timeout_work , uart_sam0_tx_timeout );
574
+ k_work_init_delayable (& dev_data -> rx_timeout_work , uart_sam0_rx_timeout );
575
575
576
576
if (cfg -> tx_dma_channel != 0xFFU ) {
577
577
struct dma_config dma_cfg = { 0 };
@@ -689,7 +689,7 @@ static void uart_sam0_isr(const struct device *dev)
689
689
if (dev_data -> rx_timeout_time != SYS_FOREVER_MS ) {
690
690
dev_data -> rx_timeout_from_isr = true;
691
691
dev_data -> rx_timeout_start = k_uptime_get_32 ();
692
- k_delayed_work_submit (& dev_data -> rx_timeout_work ,
692
+ k_work_reschedule (& dev_data -> rx_timeout_work ,
693
693
K_MSEC (dev_data -> rx_timeout_chunk ));
694
694
}
695
695
@@ -846,7 +846,7 @@ static int uart_sam0_tx(const struct device *dev, const uint8_t *buf,
846
846
}
847
847
848
848
if (timeout != SYS_FOREVER_MS ) {
849
- k_delayed_work_submit (& dev_data -> tx_timeout_work ,
849
+ k_work_reschedule (& dev_data -> tx_timeout_work ,
850
850
K_MSEC (timeout ));
851
851
}
852
852
@@ -865,7 +865,7 @@ static int uart_sam0_tx_abort(const struct device *dev)
865
865
return - ENOTSUP ;
866
866
}
867
867
868
- k_delayed_work_cancel (& dev_data -> tx_timeout_work );
868
+ k_work_cancel_delayable (& dev_data -> tx_timeout_work );
869
869
870
870
return uart_sam0_tx_halt (dev_data );
871
871
}
@@ -965,7 +965,7 @@ static int uart_sam0_rx_disable(const struct device *dev)
965
965
SercomUsart * const regs = cfg -> regs ;
966
966
struct dma_status st ;
967
967
968
- k_delayed_work_cancel (& dev_data -> rx_timeout_work );
968
+ k_work_cancel_delayable (& dev_data -> rx_timeout_work );
969
969
970
970
int key = irq_lock ();
971
971
0 commit comments