rt-thread/bsp/at32/libraries/rt_drivers/drv_emac.c

1071 lines
29 KiB
C
Raw Normal View History

2022-04-08 15:31:35 +08:00
/*
* Copyright (c) 2006-2021, RT-Thread Development Team
*
* SPDX-License-Identifier: Apache-2.0
*
* Change Logs:
* Date Author Notes
* 2022-05-16 shelton first version
* 2022-07-11 shelton optimize code to improve network throughput
* performance
2022-10-14 10:03:27 +08:00
* 2022-10-15 shelton optimize code
* 2023-10-18 shelton optimize code
2022-04-08 15:31:35 +08:00
*/
#include "drv_emac.h"
#include <netif/ethernetif.h>
#include <lwipopts.h>
/* debug option */
//#define EMAC_RX_DUMP
//#define EMAC_TX_DUMP
//#define DRV_DEBUG
#define LOG_TAG "drv.emac"
#include <drv_log.h>
#define CRYSTAL_ON_PHY 0
/* emac memory buffer configuration */
#define EMAC_NUM_RX_BUF 5 /* rx (5 * 1500) */
#define EMAC_NUM_TX_BUF 5 /* tx (5 * 1500) */
2022-04-08 15:31:35 +08:00
#define MAX_ADDR_LEN 6
#define DMARXDESC_FRAMELENGTH_SHIFT 16
2022-04-08 15:31:35 +08:00
struct rt_at32_emac
{
/* inherit from ethernet device */
struct eth_device parent;
#ifndef PHY_USING_INTERRUPT_MODE
rt_timer_t poll_link_timer;
#endif
/* interface address info, hw address */
rt_uint8_t dev_addr[MAX_ADDR_LEN];
/* emac_speed */
emac_speed_type emac_speed;
/* emac_duplex_mode */
emac_duplex_type emac_mode;
};
typedef struct {
rt_uint32_t length;
rt_uint32_t buffer;
emac_dma_desc_type *descriptor;
emac_dma_desc_type *rx_fs_desc;
emac_dma_desc_type *rx_ls_desc;
rt_uint8_t g_seg_count;
} frame_type;
2022-04-08 15:31:35 +08:00
static emac_dma_desc_type *dma_rx_dscr_tab, *dma_tx_dscr_tab;
extern emac_dma_desc_type *dma_rx_desc_to_get, *dma_tx_desc_to_set;
frame_type rx_frame;
2022-04-08 15:31:35 +08:00
static rt_uint8_t *rx_buff, *tx_buff;
static struct rt_at32_emac at32_emac_device;
static uint8_t phy_addr = 0xFF;
#if defined(EMAC_RX_DUMP) || defined(EMAC_TX_DUMP)
#define __is_print(ch) ((unsigned int)((ch) - ' ') < 127u - ' ')
static void dump_hex(const rt_uint8_t *ptr, rt_size_t buflen)
{
unsigned char *buf = (unsigned char *)ptr;
int i, j;
for (i = 0; i < buflen; i += 16)
{
rt_kprintf("%08X: ", i);
for (j = 0; j < 16; j++)
if (i + j < buflen)
rt_kprintf("%02X ", buf[i + j]);
else
rt_kprintf(" ");
rt_kprintf(" ");
for (j = 0; j < 16; j++)
if (i + j < buflen)
rt_kprintf("%c", __is_print(buf[i + j]) ? buf[i + j] : '.');
rt_kprintf("\n");
}
}
#endif
/**
* @brief phy reset
*/
static void phy_reset(void)
{
gpio_init_type gpio_init_struct;
#if defined (SOC_SERIES_AT32F437)
2022-04-08 15:31:35 +08:00
crm_periph_clock_enable(CRM_GPIOE_PERIPH_CLOCK, TRUE);
crm_periph_clock_enable(CRM_GPIOG_PERIPH_CLOCK, TRUE);
gpio_default_para_init(&gpio_init_struct);
gpio_init_struct.gpio_drive_strength = GPIO_DRIVE_STRENGTH_STRONGER;
gpio_init_struct.gpio_mode = GPIO_MODE_OUTPUT;
gpio_init_struct.gpio_out_type = GPIO_OUTPUT_PUSH_PULL;
gpio_init_struct.gpio_pull = GPIO_PULL_NONE;
gpio_init_struct.gpio_pins = GPIO_PINS_15;
gpio_init(GPIOE, &gpio_init_struct);
gpio_init_struct.gpio_pins = GPIO_PINS_15;
gpio_init(GPIOG, &gpio_init_struct);
gpio_bits_reset(GPIOE, GPIO_PINS_15);
gpio_bits_reset(GPIOG, GPIO_PINS_15);
rt_thread_mdelay(2);
gpio_bits_set(GPIOE, GPIO_PINS_15);
#endif
#if defined (SOC_SERIES_AT32F407)
crm_periph_clock_enable(CRM_GPIOC_PERIPH_CLOCK, TRUE);
gpio_default_para_init(&gpio_init_struct);
gpio_init_struct.gpio_drive_strength = GPIO_DRIVE_STRENGTH_STRONGER;
gpio_init_struct.gpio_mode = GPIO_MODE_OUTPUT;
gpio_init_struct.gpio_out_type = GPIO_OUTPUT_PUSH_PULL;
gpio_init_struct.gpio_pull = GPIO_PULL_NONE;
gpio_init_struct.gpio_pins = GPIO_PINS_8;
gpio_init(GPIOC, &gpio_init_struct);
gpio_bits_reset(GPIOC, GPIO_PINS_8);
rt_thread_mdelay(2);
gpio_bits_set(GPIOC, GPIO_PINS_8);
#endif
2022-04-08 15:31:35 +08:00
rt_thread_mdelay(2000);
}
/**
* @brief phy clock config
*/
static void phy_clock_config(void)
{
#if (CRYSTAL_ON_PHY == 0)
/* if CRYSTAL_NO_PHY, output clock with pa8 of mcu */
gpio_init_type gpio_init_struct;
crm_periph_clock_enable(CRM_GPIOA_PERIPH_CLOCK, TRUE);
gpio_default_para_init(&gpio_init_struct);
gpio_init_struct.gpio_drive_strength = GPIO_DRIVE_STRENGTH_STRONGER;
gpio_init_struct.gpio_mode = GPIO_MODE_MUX;
gpio_init_struct.gpio_out_type = GPIO_OUTPUT_PUSH_PULL;
gpio_init_struct.gpio_pull = GPIO_PULL_NONE;
gpio_init_struct.gpio_pins = GPIO_PINS_8;
gpio_init(GPIOA, &gpio_init_struct);
/* 9162 clkout output 25 mhz */
/* 83848 clkout output 50 mhz */
#if defined (SOC_SERIES_AT32F407)
crm_clock_out_set(CRM_CLKOUT_SCLK);
#if defined (PHY_USING_DM9162)
crm_clkout_div_set(CRM_CLKOUT_DIV_8);
#elif defined (PHY_USING_DP83848)
crm_clkout_div_set(CRM_CLKOUT_DIV_4);
#endif
#endif
#if defined (SOC_SERIES_AT32F437)
2022-04-08 15:31:35 +08:00
crm_clock_out1_set(CRM_CLKOUT1_PLL);
#if defined (PHY_USING_DM9162)
crm_clkout_div_set(CRM_CLKOUT_INDEX_1, CRM_CLKOUT_DIV1_5, CRM_CLKOUT_DIV2_2);
#elif defined (PHY_USING_DP83848)
crm_clkout_div_set(CRM_CLKOUT_INDEX_1, CRM_CLKOUT_DIV1_5, CRM_CLKOUT_DIV2_1);
#endif
#endif
#endif
2022-04-08 15:31:35 +08:00
}
/**
* @brief reset phy register
*/
static error_status emac_phy_register_reset(void)
{
uint16_t data = 0;
uint32_t timeout = 0;
uint32_t i = 0;
if(emac_phy_register_write(phy_addr, PHY_CONTROL_REG, PHY_RESET_BIT) == ERROR)
{
return ERROR;
}
for(i = 0; i < 0x000FFFFF; i++);
do
{
timeout++;
if(emac_phy_register_read(phy_addr, PHY_CONTROL_REG, &data) == ERROR)
{
return ERROR;
}
} while((data & PHY_RESET_BIT) && (timeout < PHY_TIMEOUT));
for(i = 0; i < 0x00FFFFF; i++);
if(timeout == PHY_TIMEOUT)
{
return ERROR;
}
return SUCCESS;
}
/**
* @brief set mac speed related parameters
*/
static error_status emac_speed_config(emac_auto_negotiation_type nego, emac_duplex_type mode, emac_speed_type speed)
{
uint16_t data = 0;
uint32_t timeout = 0;
if(nego == EMAC_AUTO_NEGOTIATION_ON)
{
do
{
timeout++;
if(emac_phy_register_read(phy_addr, PHY_STATUS_REG, &data) == ERROR)
{
return ERROR;
}
} while(!(data & PHY_LINKED_STATUS_BIT) && (timeout < PHY_TIMEOUT));
if(timeout == PHY_TIMEOUT)
{
return ERROR;
}
timeout = 0;
if(emac_phy_register_write(phy_addr, PHY_CONTROL_REG, PHY_AUTO_NEGOTIATION_BIT) == ERROR)
{
return ERROR;
}
do
{
timeout++;
if(emac_phy_register_read(phy_addr, PHY_STATUS_REG, &data) == ERROR)
{
return ERROR;
}
} while(!(data & PHY_NEGO_COMPLETE_BIT) && (timeout < PHY_TIMEOUT));
if(timeout == PHY_TIMEOUT)
{
return ERROR;
}
if(emac_phy_register_read(phy_addr, PHY_SPECIFIED_CS_REG, &data) == ERROR)
{
return ERROR;
}
#ifdef PHY_USING_DM9162
if(data & PHY_FULL_DUPLEX_100MBPS_BIT)
{
emac_fast_speed_set(EMAC_SPEED_100MBPS);
emac_duplex_mode_set(EMAC_FULL_DUPLEX);
}
else if(data & PHY_HALF_DUPLEX_100MBPS_BIT)
{
emac_fast_speed_set(EMAC_SPEED_100MBPS);
emac_duplex_mode_set(EMAC_HALF_DUPLEX);
}
else if(data & PHY_FULL_DUPLEX_10MBPS_BIT)
{
emac_fast_speed_set(EMAC_SPEED_10MBPS);
emac_duplex_mode_set(EMAC_FULL_DUPLEX);
}
else if(data & PHY_HALF_DUPLEX_10MBPS_BIT)
{
emac_fast_speed_set(EMAC_SPEED_10MBPS);
emac_duplex_mode_set(EMAC_HALF_DUPLEX);
}
#endif
#ifdef PHY_USING_DP83848
if(data & PHY_DUPLEX_MODE)
{
emac_duplex_mode_set(EMAC_FULL_DUPLEX);
}
else
{
emac_duplex_mode_set(EMAC_HALF_DUPLEX);
}
if(data & PHY_SPEED_MODE)
{
emac_fast_speed_set(EMAC_SPEED_10MBPS);
}
else
{
emac_fast_speed_set(EMAC_SPEED_100MBPS);
}
#endif
}
else
{
if(emac_phy_register_write(phy_addr, PHY_CONTROL_REG, (uint16_t)((mode << 8) | (speed << 13))) == ERROR)
{
return ERROR;
}
if(speed == EMAC_SPEED_100MBPS)
{
emac_fast_speed_set(EMAC_SPEED_100MBPS);
}
else
{
emac_fast_speed_set(EMAC_SPEED_10MBPS);
}
if(mode == EMAC_FULL_DUPLEX)
{
emac_duplex_mode_set(EMAC_FULL_DUPLEX);
}
else
{
emac_duplex_mode_set(EMAC_HALF_DUPLEX);
}
}
return SUCCESS;
}
/**
* @brief initialize emac phy
*/
static error_status emac_phy_init(emac_control_config_type *control_para)
{
emac_clock_range_set();
if(emac_phy_register_reset() == ERROR)
{
return ERROR;
}
if(emac_speed_config(control_para->auto_nego, control_para->duplex_mode, control_para->fast_ethernet_speed) == ERROR)
{
return ERROR;
}
emac_control_config(control_para);
return SUCCESS;
}
/**
* @brief emac initialization function
*/
static rt_err_t rt_at32_emac_init(rt_device_t dev)
{
emac_control_config_type mac_control_para;
emac_dma_config_type dma_control_para;
/* check till phy detected */
while(phy_addr == 0xFF)
{
rt_thread_mdelay(1000);
}
/* emac reset */
emac_reset();
/* software reset emac dma */
emac_dma_software_reset_set();
while(emac_dma_software_reset_get() == SET);
emac_control_para_init(&mac_control_para);
mac_control_para.auto_nego = EMAC_AUTO_NEGOTIATION_ON;
2022-10-14 10:03:27 +08:00
2022-04-08 15:31:35 +08:00
if(emac_phy_init(&mac_control_para) == ERROR)
{
LOG_E("emac hardware init failed");
return -RT_ERROR;
2022-04-08 15:31:35 +08:00
}
else
{
LOG_D("emac hardware init success");
}
emac_transmit_flow_control_enable(TRUE);
emac_zero_quanta_pause_disable(TRUE);
/* set mac address */
emac_local_address_set(at32_emac_device.dev_addr);
/* set emac dma rx link list */
emac_dma_descriptor_list_address_set(EMAC_DMA_RECEIVE, dma_rx_dscr_tab, rx_buff, EMAC_NUM_RX_BUF);
/* set emac dma tx link list */
emac_dma_descriptor_list_address_set(EMAC_DMA_TRANSMIT, dma_tx_dscr_tab, tx_buff, EMAC_NUM_TX_BUF);
2022-10-14 10:03:27 +08:00
emac_dma_para_init(&dma_control_para);
dma_control_para.rsf_enable = TRUE;
dma_control_para.tsf_enable = TRUE;
dma_control_para.osf_enable = TRUE;
dma_control_para.aab_enable = TRUE;
dma_control_para.usp_enable = TRUE;
dma_control_para.fb_enable = TRUE;
dma_control_para.flush_rx_disable = TRUE;
dma_control_para.rx_dma_pal = EMAC_DMA_PBL_32;
dma_control_para.tx_dma_pal = EMAC_DMA_PBL_32;
dma_control_para.priority_ratio = EMAC_DMA_2_RX_1_TX;
2022-04-08 15:31:35 +08:00
emac_dma_config(&dma_control_para);
2022-10-14 10:03:27 +08:00
/* emac interrupt init */
2022-04-08 15:31:35 +08:00
emac_dma_interrupt_enable(EMAC_DMA_INTERRUPT_NORMAL_SUMMARY, TRUE);
emac_dma_interrupt_enable(EMAC_DMA_INTERRUPT_RX, TRUE);
nvic_irq_enable(EMAC_IRQn, 0x07, 0);
/* enable emac */
emac_start();
return RT_EOK;
}
static rt_err_t rt_at32_emac_open(rt_device_t dev, rt_uint16_t oflag)
{
LOG_D("emac open");
return RT_EOK;
}
static rt_err_t rt_at32_emac_close(rt_device_t dev)
{
LOG_D("emac close");
return RT_EOK;
}
static rt_ssize_t rt_at32_emac_read(rt_device_t dev, rt_off_t pos, void *buffer, rt_size_t size)
2022-04-08 15:31:35 +08:00
{
LOG_D("emac read");
rt_set_errno(-RT_ENOSYS);
return 0;
}
static rt_ssize_t rt_at32_emac_write(rt_device_t dev, rt_off_t pos, const void *buffer, rt_size_t size)
2022-04-08 15:31:35 +08:00
{
LOG_D("emac write");
rt_set_errno(-RT_ENOSYS);
return 0;
}
static rt_err_t rt_at32_emac_control(rt_device_t dev, int cmd, void *args)
{
switch (cmd)
{
case NIOCTL_GADDR:
/* get mac address */
if (args)
{
SMEMCPY(args, at32_emac_device.dev_addr, 6);
}
else
{
return -RT_ERROR;
}
break;
default :
break;
}
return RT_EOK;
}
/**
* @brief emac txpkt chainmode
2022-04-08 15:31:35 +08:00
*/
rt_err_t emac_txpkt_chainmode(rt_uint32_t frame_length)
2022-04-08 15:31:35 +08:00
{
rt_uint32_t buf_cnt = 0, index = 0;
2022-04-08 15:31:35 +08:00
/* check if the descriptor is owned by the ethernet dma (when set) or cpu (when reset) */
if((dma_tx_desc_to_set->status & EMAC_DMATXDESC_OWN) != (u32)RESET)
2022-04-08 15:31:35 +08:00
{
/* return error: own bit set */
return RT_ERROR;
2022-04-08 15:31:35 +08:00
}
if(frame_length == 0)
2022-04-08 15:31:35 +08:00
{
return RT_ERROR;
}
2022-04-08 15:31:35 +08:00
if(frame_length > EMAC_MAX_PACKET_LENGTH)
{
buf_cnt = frame_length / EMAC_MAX_PACKET_LENGTH;
if(frame_length % EMAC_MAX_PACKET_LENGTH)
{
buf_cnt += 1;
}
}
else
{
buf_cnt = 1;
2022-04-08 15:31:35 +08:00
}
if(buf_cnt == 1)
{
/* setting the last segment and first segment bits (in this case a frame is transmitted in one descriptor) */
dma_tx_desc_to_set->status |= EMAC_DMATXDESC_LS | EMAC_DMATXDESC_FS;
/* setting the frame length: bits[12:0] */
dma_tx_desc_to_set->controlsize = (frame_length & EMAC_DMATXDESC_TBS1);
/* set own bit of the tx descriptor status: gives the buffer back to ethernet dma */
dma_tx_desc_to_set->status |= EMAC_DMATXDESC_OWN;
/* selects the next dma tx descriptor list for next buffer to send */
dma_tx_desc_to_set = (emac_dma_desc_type*) (dma_tx_desc_to_set->buf2nextdescaddr);
}
else
{
for(index = 0; index < buf_cnt; index ++)
{
/* clear first and last segments */
dma_tx_desc_to_set->status &= ~(EMAC_DMATXDESC_LS | EMAC_DMATXDESC_FS);
/* set first segments */
if(index == 0)
{
dma_tx_desc_to_set->status |= EMAC_DMATXDESC_FS;
}
/* set size */
dma_tx_desc_to_set->controlsize = (EMAC_MAX_PACKET_LENGTH & EMAC_DMATXDESC_TBS1);
2022-10-14 10:03:27 +08:00
/* set last segments */
if(index == (buf_cnt - 1))
{
dma_tx_desc_to_set->status |= EMAC_DMATXDESC_LS;
dma_tx_desc_to_set->controlsize = ((frame_length - ((buf_cnt-1) * EMAC_MAX_PACKET_LENGTH)) & EMAC_DMATXDESC_TBS1);
}
2022-04-08 15:31:35 +08:00
/* set own bit of the tx descriptor status: gives the buffer back to ethernet dma */
dma_tx_desc_to_set->status |= EMAC_DMATXDESC_OWN;
2022-04-08 15:31:35 +08:00
/* selects the next dma tx descriptor list for next buffer to send */
dma_tx_desc_to_set = (emac_dma_desc_type*) (dma_tx_desc_to_set->buf2nextdescaddr);
}
}
/* when tx buffer unavailable flag is set: clear it and resume transmission */
if(emac_dma_flag_get(EMAC_DMA_TBU_FLAG))
2022-04-08 15:31:35 +08:00
{
/* clear tbus ethernet dma flag */
2022-04-08 15:31:35 +08:00
emac_dma_flag_clear(EMAC_DMA_TBU_FLAG);
/* resume dma transmission*/
EMAC_DMA->tpd_bit.tpd = 0;
2022-04-08 15:31:35 +08:00
}
return RT_EOK;
}
2022-04-08 15:31:35 +08:00
/**
* @brief transmit data
*/
rt_err_t rt_at32_emac_tx(rt_device_t dev, struct pbuf *p)
{
rt_err_t ret = -RT_ERROR;
struct pbuf *q;
rt_uint32_t length = 0;
rt_uint32_t buffer_offset = 0, payload_offset = 0, copy_count = 0;
emac_dma_desc_type *dma_tx_desc;
rt_uint8_t *buffer;
dma_tx_desc = dma_tx_desc_to_set;
buffer = (uint8_t *)(dma_tx_desc_to_set->buf1addr);
/* copy data to buffer */
for(q = p; q != NULL; q = q->next)
{
if((dma_tx_desc->status & EMAC_DMATXDESC_OWN) != RESET)
{
ret = RT_EOK;
goto _error;
}
copy_count = q->len;
payload_offset = 0;
while((copy_count + buffer_offset) > EMAC_MAX_PACKET_LENGTH)
{
rt_memcpy(buffer + buffer_offset, (uint8_t *)q->payload + payload_offset, (EMAC_MAX_PACKET_LENGTH - buffer_offset));
dma_tx_desc = (emac_dma_desc_type*)dma_tx_desc->buf2nextdescaddr;
if((dma_tx_desc->status & EMAC_DMATXDESC_OWN) != RESET)
{
ret = RT_EOK;
goto _error;
}
buffer = (uint8_t *)dma_tx_desc->buf1addr;
copy_count = copy_count - (EMAC_MAX_PACKET_LENGTH - buffer_offset);
payload_offset = payload_offset + (EMAC_MAX_PACKET_LENGTH - buffer_offset);
length = length + (EMAC_MAX_PACKET_LENGTH - buffer_offset);
buffer_offset = 0;
}
rt_memcpy(buffer + buffer_offset, (uint8_t *)q->payload + payload_offset, copy_count);
buffer_offset = buffer_offset + copy_count;
length = length + copy_count;
}
emac_txpkt_chainmode(length);
ret = RT_EOK;
_error:
/* when tx buffer unavailable flag is set: clear it and resume transmission */
if(emac_dma_flag_get(EMAC_DMA_TBU_FLAG))
{
/* clear tbus ethernet dma flag */
emac_dma_flag_clear(EMAC_DMA_TBU_FLAG);
/* resume dma transmission*/
EMAC_DMA->tpd_bit.tpd = 0;
}
return ret;
2022-04-08 15:31:35 +08:00
}
/**
* @brief emac rxpkt chainmode
*/
rt_err_t emac_rxpkt_chainmode(void)
{
/* check if the descriptor is owned by the ethernet dma (when set) or cpu (when reset) */
if((dma_rx_desc_to_get->status & EMAC_DMARXDESC_OWN) != (u32)RESET)
{
/* return error: own bit set */
return RT_ERROR;
}
if((dma_rx_desc_to_get->status & EMAC_DMARXDESC_LS) != (u32)RESET)
{
rx_frame.g_seg_count ++;
if(rx_frame.g_seg_count == 1)
{
rx_frame.rx_fs_desc = dma_rx_desc_to_get;
}
rx_frame.rx_ls_desc = dma_rx_desc_to_get;
rx_frame.length = ((dma_rx_desc_to_get->status & EMAC_DMARXDESC_FL) >> DMARXDESC_FRAMELENGTH_SHIFT) - 4;
rx_frame.buffer = rx_frame.rx_fs_desc->buf1addr;
/* Selects the next DMA Rx descriptor list for next buffer to read */
dma_rx_desc_to_get = (emac_dma_desc_type*) (dma_rx_desc_to_get->buf2nextdescaddr);
return RT_EOK;
}
else if((dma_rx_desc_to_get->status & EMAC_DMARXDESC_FS) != (u32)RESET)
{
rx_frame.g_seg_count = 1;
rx_frame.rx_fs_desc = dma_rx_desc_to_get;
rx_frame.rx_ls_desc = NULL;
dma_rx_desc_to_get = (emac_dma_desc_type*) (dma_rx_desc_to_get->buf2nextdescaddr);
}
else
{
rx_frame.g_seg_count ++;
dma_rx_desc_to_get = (emac_dma_desc_type*) (dma_rx_desc_to_get->buf2nextdescaddr);
}
return RT_ERROR;
}
2022-04-08 15:31:35 +08:00
/**
* @brief receive data
*/
struct pbuf *rt_at32_emac_rx(rt_device_t dev)
{
struct pbuf *p = NULL;
struct pbuf *q = NULL;
rt_uint16_t len = 0;
rt_uint8_t *buffer;
2022-04-08 15:31:35 +08:00
emac_dma_desc_type *dma_rx_desc;
rt_uint32_t buffer_offset, payload_offset = 0, copy_count = 0;
rt_uint32_t index = 0;
if(emac_rxpkt_chainmode() != RT_EOK)
{
return NULL;
}
/* obtain the size of the packet and put it into the "len"
variable. */
len = rx_frame.length;
buffer = (uint8_t *)rx_frame.buffer;
/* we allocate a pbuf chain of pbufs from the pool. */
2022-04-08 15:31:35 +08:00
if(len > 0)
{
p = pbuf_alloc(PBUF_RAW, len, PBUF_POOL);
}
2022-04-08 15:31:35 +08:00
if(p != NULL)
{
dma_rx_desc = rx_frame.rx_fs_desc;
buffer_offset = 0;
for (q = p; q != NULL; q = q->next)
2022-04-08 15:31:35 +08:00
{
copy_count = q->len;
payload_offset = 0;
while( (copy_count + buffer_offset) > EMAC_MAX_PACKET_LENGTH )
2022-04-08 15:31:35 +08:00
{
/* copy data to pbuf */
rt_memcpy((uint8_t*)q->payload + payload_offset, buffer + buffer_offset, (EMAC_MAX_PACKET_LENGTH - buffer_offset));
/* point to next descriptor */
dma_rx_desc = (emac_dma_desc_type *)(dma_rx_desc->buf2nextdescaddr);
buffer = (uint8_t *)(dma_rx_desc->buf1addr);
copy_count = copy_count - (EMAC_MAX_PACKET_LENGTH - buffer_offset);
payload_offset = payload_offset + (EMAC_MAX_PACKET_LENGTH - buffer_offset);
buffer_offset = 0;
2022-04-08 15:31:35 +08:00
}
rt_memcpy((uint8_t*)q->payload + payload_offset, (uint8_t*)buffer + buffer_offset, copy_count);
buffer_offset = buffer_offset + copy_count;
2022-04-08 15:31:35 +08:00
}
}
dma_rx_desc = rx_frame.rx_fs_desc;
for(index = 0; index < rx_frame.g_seg_count; index ++)
2022-04-08 15:31:35 +08:00
{
dma_rx_desc->status |= EMAC_DMARXDESC_OWN;
dma_rx_desc = (emac_dma_desc_type*) (dma_rx_desc->buf2nextdescaddr);
2022-04-08 15:31:35 +08:00
}
rx_frame.g_seg_count = 0;
2022-04-08 15:31:35 +08:00
/* when rx buffer unavailable flag is set: clear it and resume reception */
if(emac_dma_flag_get(EMAC_DMA_RBU_FLAG))
2022-04-08 15:31:35 +08:00
{
/* clear rbus ethernet dma flag */
2022-04-08 15:31:35 +08:00
emac_dma_flag_clear(EMAC_DMA_RBU_FLAG);
/* resume dma reception */
EMAC_DMA->rpd_bit.rpd = FALSE;
2022-04-08 15:31:35 +08:00
}
return p;
}
void EMAC_IRQHandler(void)
{
/* enter interrupt */
rt_interrupt_enter();
/* packet receiption */
if (emac_dma_flag_get(EMAC_DMA_RI_FLAG) == SET)
{
/* a frame has been received */
eth_device_ready(&(at32_emac_device.parent));
emac_dma_flag_clear(EMAC_DMA_RI_FLAG);
}
/* packet transmission */
if (emac_dma_flag_get(EMAC_DMA_TI_FLAG) == SET)
{
emac_dma_flag_clear(EMAC_DMA_TI_FLAG);
}
/* clear normal interrupt */
emac_dma_flag_clear(EMAC_DMA_NIS_FLAG);
/* clear dma error */
if(emac_dma_flag_get(EMAC_DMA_AIS_FLAG) != RESET)
{
if(emac_dma_flag_get(EMAC_DMA_RBU_FLAG) != RESET)
2022-04-08 15:31:35 +08:00
{
emac_dma_flag_clear(EMAC_DMA_RBU_FLAG);
}
if(emac_dma_flag_get(EMAC_DMA_OVF_FLAG) != RESET)
{
emac_dma_flag_clear(EMAC_DMA_OVF_FLAG);
2022-04-08 15:31:35 +08:00
}
emac_dma_flag_clear(EMAC_DMA_AIS_FLAG);
2022-04-08 15:31:35 +08:00
}
/* leave interrupt */
rt_interrupt_leave();
}
enum {
PHY_LINK = (1 << 0),
PHY_10M = (1 << 1),
PHY_FULLDUPLEX = (1 << 2),
};
static void phy_linkchange()
{
static rt_uint8_t phy_speed = 0;
rt_uint8_t phy_speed_new = 0;
rt_uint16_t status;
emac_phy_register_read(phy_addr, PHY_BASIC_STATUS_REG, (uint16_t *)&status);
LOG_D("phy basic status reg is 0x%X", status);
if (status & (PHY_AUTONEGO_COMPLETE_MASK | PHY_LINKED_STATUS_MASK))
{
rt_uint16_t SR = 0;
phy_speed_new |= PHY_LINK;
emac_phy_register_read(phy_addr, PHY_SPECIFIED_CS_REG, (uint16_t *)&SR);
LOG_D("phy control status reg is 0x%X", SR);
if (SR & (PHY_SPEED_MODE))
{
phy_speed_new |= PHY_10M;
}
if (SR & (PHY_DUPLEX_MODE))
{
phy_speed_new |= PHY_FULLDUPLEX;
}
}
if (phy_speed != phy_speed_new)
{
phy_speed = phy_speed_new;
if (phy_speed & PHY_LINK)
{
LOG_D("link up");
if (phy_speed & PHY_10M)
{
LOG_D("10Mbps");
at32_emac_device.emac_speed = EMAC_SPEED_10MBPS;
}
else
{
at32_emac_device.emac_speed = EMAC_SPEED_100MBPS;
LOG_D("100Mbps");
}
if (phy_speed & PHY_FULLDUPLEX)
{
LOG_D("full-duplex");
at32_emac_device.emac_mode = EMAC_FULL_DUPLEX;
}
else
{
LOG_D("half-duplex");
at32_emac_device.emac_mode = EMAC_HALF_DUPLEX;
}
/* send link up. */
eth_device_linkchange(&at32_emac_device.parent, RT_TRUE);
}
else
{
LOG_I("link down");
eth_device_linkchange(&at32_emac_device.parent, RT_FALSE);
}
}
}
#ifdef PHY_USING_INTERRUPT_MODE
static void emac_phy_isr(void *args)
{
rt_uint32_t status = 0;
emac_phy_register_read(phy_addr, PHY_INTERRUPT_FLAG_REG, (uint16_t *)&status);
LOG_D("phy interrupt status reg is 0x%X", status);
phy_linkchange();
}
#endif /* PHY_USING_INTERRUPT_MODE */
static void phy_monitor_thread_entry(void *parameter)
{
uint8_t detected_count = 0;
while(phy_addr == 0xFF)
{
/* phy search */
rt_uint32_t i, temp;
for (i = 0; i <= 0x1F; i++)
{
emac_phy_register_read(i, PHY_BASIC_STATUS_REG, (uint16_t *)&temp);
if (temp != 0xFFFF && temp != 0x00)
{
phy_addr = i;
break;
}
}
detected_count++;
rt_thread_mdelay(1000);
if (detected_count > 10)
{
LOG_E("No PHY device was detected, please check hardware!");
}
}
LOG_D("Found a phy, address:0x%02X", phy_addr);
/* reset phy */
LOG_D("RESET PHY!");
emac_phy_register_write(phy_addr, PHY_BASIC_CONTROL_REG, PHY_RESET_MASK);
rt_thread_mdelay(2000);
emac_phy_register_write(phy_addr, PHY_BASIC_CONTROL_REG, PHY_AUTO_NEGOTIATION_MASK);
phy_linkchange();
#ifdef PHY_USING_INTERRUPT_MODE
/* configuration intterrupt pin */
rt_pin_mode(PHY_INT_PIN, PIN_MODE_INPUT_PULLUP);
rt_pin_attach_irq(PHY_INT_PIN, PIN_IRQ_MODE_FALLING, emac_phy_isr, (void *)"callbackargs");
rt_pin_irq_enable(PHY_INT_PIN, PIN_IRQ_ENABLE);
/* enable phy interrupt */
emac_phy_register_write(phy_addr, PHY_INTERRUPT_MASK_REG, PHY_INT_MASK);
#if defined(PHY_INTERRUPT_CTRL_REG)
emac_phy_register_write(phy_addr, PHY_INTERRUPT_CTRL_REG, PHY_INTERRUPT_EN);
#endif
#else /* PHY_USING_INTERRUPT_MODE */
at32_emac_device.poll_link_timer = rt_timer_create("phylnk", (void (*)(void*))phy_linkchange,
NULL, RT_TICK_PER_SECOND, RT_TIMER_FLAG_PERIODIC);
if (!at32_emac_device.poll_link_timer || rt_timer_start(at32_emac_device.poll_link_timer) != RT_EOK)
{
LOG_E("Start link change detection timer failed");
}
#endif /* PHY_USING_INTERRUPT_MODE */
}
/* Register the EMAC device */
static int rt_hw_at32_emac_init(void)
{
rt_err_t state = RT_EOK;
/* Prepare receive and send buffers */
rx_buff = (rt_uint8_t *)rt_calloc(EMAC_NUM_RX_BUF, EMAC_MAX_PACKET_LENGTH);
if (rx_buff == RT_NULL)
{
LOG_E("No memory");
state = -RT_ENOMEM;
goto __exit;
}
tx_buff = (rt_uint8_t *)rt_calloc(EMAC_NUM_TX_BUF, EMAC_MAX_PACKET_LENGTH);
if (tx_buff == RT_NULL)
{
LOG_E("No memory");
state = -RT_ENOMEM;
goto __exit;
}
dma_rx_dscr_tab = (emac_dma_desc_type *)rt_calloc(EMAC_NUM_RX_BUF, sizeof(emac_dma_desc_type));
if (dma_rx_dscr_tab == RT_NULL)
{
LOG_E("No memory");
state = -RT_ENOMEM;
goto __exit;
}
dma_tx_dscr_tab = (emac_dma_desc_type *)rt_calloc(EMAC_NUM_TX_BUF, sizeof(emac_dma_desc_type));
if (dma_tx_dscr_tab == RT_NULL)
{
LOG_E("No memory");
state = -RT_ENOMEM;
goto __exit;
}
/* phy clock */
phy_clock_config();
/* enable periph clock */
crm_periph_clock_enable(CRM_EMAC_PERIPH_CLOCK, TRUE);
crm_periph_clock_enable(CRM_EMACTX_PERIPH_CLOCK, TRUE);
crm_periph_clock_enable(CRM_EMACRX_PERIPH_CLOCK, TRUE);
/* interface mode */
#if defined (SOC_SERIES_AT32F407)
gpio_pin_remap_config(MII_RMII_SEL_GMUX, TRUE);
#endif
#if defined (SOC_SERIES_AT32F437)
2022-04-08 15:31:35 +08:00
scfg_emac_interface_set(SCFG_EMAC_SELECT_RMII);
#endif
2022-04-08 15:31:35 +08:00
/* emac gpio init */
at32_msp_emac_init(NULL);
at32_emac_device.emac_speed = EMAC_SPEED_100MBPS;
at32_emac_device.emac_mode = EMAC_FULL_DUPLEX;
at32_emac_device.dev_addr[0] = 0x00;
at32_emac_device.dev_addr[1] = 0x66;
at32_emac_device.dev_addr[2] = 0x88;
2022-04-08 15:31:35 +08:00
/* generate mac addr from unique id (only for test). */
at32_emac_device.dev_addr[3] = *(rt_uint8_t *)(0x1FFFF7E8 + 4);
at32_emac_device.dev_addr[4] = *(rt_uint8_t *)(0x1FFFF7E8 + 2);
at32_emac_device.dev_addr[5] = *(rt_uint8_t *)(0x1FFFF7E8 + 0);
at32_emac_device.parent.parent.init = rt_at32_emac_init;
at32_emac_device.parent.parent.open = rt_at32_emac_open;
at32_emac_device.parent.parent.close = rt_at32_emac_close;
at32_emac_device.parent.parent.read = rt_at32_emac_read;
at32_emac_device.parent.parent.write = rt_at32_emac_write;
at32_emac_device.parent.parent.control = rt_at32_emac_control;
at32_emac_device.parent.parent.user_data = RT_NULL;
at32_emac_device.parent.eth_rx = rt_at32_emac_rx;
at32_emac_device.parent.eth_tx = rt_at32_emac_tx;
rx_frame.g_seg_count = 0;
2022-04-08 15:31:35 +08:00
/* reset phy */
phy_reset();
/* start phy monitor */
rt_thread_t tid;
tid = rt_thread_create("phy",
phy_monitor_thread_entry,
RT_NULL,
1024,
RT_THREAD_PRIORITY_MAX - 2,
2);
if (tid != RT_NULL)
{
rt_thread_startup(tid);
}
else
{
state = -RT_ERROR;
}
/* register eth device */
state = eth_device_init(&(at32_emac_device.parent), "e0");
if (RT_EOK == state)
{
LOG_D("emac device init success");
}
else
{
LOG_E("emac device init faild: %d", state);
state = -RT_ERROR;
goto __exit;
}
__exit:
if (state != RT_EOK)
{
if (rx_buff)
{
rt_free(rx_buff);
}
if (tx_buff)
{
rt_free(tx_buff);
}
if (dma_rx_dscr_tab)
{
rt_free(dma_rx_dscr_tab);
}
if (dma_tx_dscr_tab)
{
rt_free(dma_tx_dscr_tab);
}
}
return state;
}
INIT_DEVICE_EXPORT(rt_hw_at32_emac_init);