4825 lines
136 KiB
Diff
4825 lines
136 KiB
Diff
.../devicetree/bindings/net/apm-xgene-enet.txt | 72 ++
|
|
.../devicetree/bindings/pci/xgene-pci.txt | 52 ++
|
|
MAINTAINERS | 15 +
|
|
arch/arm64/Kconfig | 19 +-
|
|
arch/arm64/boot/dts/apm-mustang.dts | 12 +
|
|
arch/arm64/boot/dts/apm-storm.dtsi | 208 ++++-
|
|
arch/arm64/include/asm/Kbuild | 1 +
|
|
arch/arm64/include/asm/elf.h | 3 +-
|
|
arch/arm64/include/asm/io.h | 3 +-
|
|
arch/arm64/include/asm/kvm_mmu.h | 12 +-
|
|
arch/arm64/include/asm/pci.h | 49 ++
|
|
arch/arm64/kernel/Makefile | 1 +
|
|
arch/arm64/kernel/efi-stub.c | 16 +-
|
|
arch/arm64/kernel/efi.c | 11 +
|
|
arch/arm64/kernel/head.S | 6 +-
|
|
arch/arm64/kernel/pci.c | 38 +
|
|
arch/arm64/kernel/process.c | 6 +
|
|
arch/arm64/kernel/smp_spin_table.c | 21 +-
|
|
drivers/ata/ahci_xgene.c | 3 +-
|
|
drivers/irqchip/irq-gic.c | 32 +-
|
|
drivers/net/ethernet/Kconfig | 1 +
|
|
drivers/net/ethernet/Makefile | 1 +
|
|
drivers/net/ethernet/apm/Kconfig | 1 +
|
|
drivers/net/ethernet/apm/Makefile | 5 +
|
|
drivers/net/ethernet/apm/xgene/Kconfig | 9 +
|
|
drivers/net/ethernet/apm/xgene/Makefile | 6 +
|
|
.../net/ethernet/apm/xgene/xgene_enet_ethtool.c | 125 +++
|
|
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c | 747 ++++++++++++++++
|
|
drivers/net/ethernet/apm/xgene/xgene_enet_hw.h | 375 ++++++++
|
|
drivers/net/ethernet/apm/xgene/xgene_enet_main.c | 962 +++++++++++++++++++++
|
|
drivers/net/ethernet/apm/xgene/xgene_enet_main.h | 107 +++
|
|
drivers/of/address.c | 108 +++
|
|
drivers/of/of_pci.c | 136 +++
|
|
drivers/pci/host-bridge.c | 18 +-
|
|
drivers/pci/host/Kconfig | 10 +
|
|
drivers/pci/host/Makefile | 1 +
|
|
drivers/pci/host/pci-xgene.c | 725 ++++++++++++++++
|
|
drivers/pci/pci.c | 37 +
|
|
drivers/pci/probe.c | 68 +-
|
|
include/asm-generic/io.h | 2 +-
|
|
include/linux/of_address.h | 14 +-
|
|
include/linux/of_pci.h | 10 +
|
|
include/linux/pci.h | 15 +
|
|
tools/perf/arch/arm64/include/perf_regs.h | 2 +
|
|
44 files changed, 3992 insertions(+), 73 deletions(-)
|
|
|
|
diff --git a/Documentation/devicetree/bindings/net/apm-xgene-enet.txt b/Documentation/devicetree/bindings/net/apm-xgene-enet.txt
|
|
new file mode 100644
|
|
index 0000000..3e2a295
|
|
--- /dev/null
|
|
+++ b/Documentation/devicetree/bindings/net/apm-xgene-enet.txt
|
|
@@ -0,0 +1,72 @@
|
|
+APM X-Gene SoC Ethernet nodes
|
|
+
|
|
+Ethernet nodes are defined to describe on-chip ethernet interfaces in
|
|
+APM X-Gene SoC.
|
|
+
|
|
+Required properties:
|
|
+- compatible: Should be "apm,xgene-enet"
|
|
+- reg: Address and length of the register set for the device. It contains the
|
|
+ information of registers in the same order as described by reg-names
|
|
+- reg-names: Should contain the register set names
|
|
+ "enet_csr": Ethernet control and status register address space
|
|
+ "ring_csr": Descriptor ring control and status register address space
|
|
+ "ring_cmd": Descriptor ring command register address space
|
|
+- interrupts: Ethernet main interrupt
|
|
+- clocks: Reference to the clock entry.
|
|
+- local-mac-address: MAC address assigned to this device
|
|
+- phy-connection-type: Interface type between ethernet device and PHY device
|
|
+- phy-handle: Reference to a PHY node connected to this device
|
|
+
|
|
+- mdio: Device tree subnode with the following required
|
|
+ properties:
|
|
+
|
|
+ - compatible: Must be "apm,xgene-mdio".
|
|
+ - #address-cells: Must be <1>.
|
|
+ - #size-cells: Must be <0>.
|
|
+
|
|
+ For the phy on the mdio bus, there must be a node with the following
|
|
+ fields:
|
|
+
|
|
+ - compatible: PHY identifier. Please refer ./phy.txt for the format.
|
|
+ - reg: The ID number for the phy.
|
|
+
|
|
+Optional properties:
|
|
+- status : Should be "ok" or "disabled" for enabled/disabled.
|
|
+ Default is "ok".
|
|
+
|
|
+
|
|
+Example:
|
|
+ menetclk: menetclk {
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ clock-output-names = "menetclk";
|
|
+ status = "ok";
|
|
+ };
|
|
+
|
|
+ menet: ethernet@17020000 {
|
|
+ compatible = "apm,xgene-enet";
|
|
+ status = "disabled";
|
|
+ reg = <0x0 0x17020000 0x0 0xd100>,
|
|
+ <0x0 0X17030000 0x0 0X400>,
|
|
+ <0x0 0X10000000 0x0 0X200>;
|
|
+ reg-names = "enet_csr", "ring_csr", "ring_cmd";
|
|
+ interrupts = <0x0 0x3c 0x4>;
|
|
+ clocks = <&menetclk 0>;
|
|
+ local-mac-address = [00 01 73 00 00 01];
|
|
+ phy-connection-type = "rgmii";
|
|
+ phy-handle = <&menetphy>;
|
|
+ mdio {
|
|
+ compatible = "apm,xgene-mdio";
|
|
+ #address-cells = <1>;
|
|
+ #size-cells = <0>;
|
|
+ menetphy: menetphy@3 {
|
|
+ compatible = "ethernet-phy-id001c.c915";
|
|
+ reg = <0x3>;
|
|
+ };
|
|
+
|
|
+ };
|
|
+ };
|
|
+
|
|
+/* Board-specific peripheral configurations */
|
|
+&menet {
|
|
+ status = "ok";
|
|
+};
|
|
diff --git a/Documentation/devicetree/bindings/pci/xgene-pci.txt b/Documentation/devicetree/bindings/pci/xgene-pci.txt
|
|
new file mode 100644
|
|
index 0000000..e19fdb8
|
|
--- /dev/null
|
|
+++ b/Documentation/devicetree/bindings/pci/xgene-pci.txt
|
|
@@ -0,0 +1,52 @@
|
|
+* AppliedMicro X-Gene PCIe interface
|
|
+
|
|
+Required properties:
|
|
+- device_type: set to "pci"
|
|
+- compatible: should contain "apm,xgene-pcie" to identify the core.
|
|
+- reg: A list of physical base address and length for each set of controller
|
|
+ registers. Must contain an entry for each entry in the reg-names
|
|
+ property.
|
|
+- reg-names: Must include the following entries:
|
|
+ "csr": controller configuration registers.
|
|
+ "cfg": pcie configuration space registers.
|
|
+- #address-cells: set to <3>
|
|
+- #size-cells: set to <2>
|
|
+- ranges: ranges for the outbound memory, I/O regions.
|
|
+- dma-ranges: ranges for the inbound memory regions.
|
|
+- #interrupt-cells: set to <1>
|
|
+- interrupt-map-mask and interrupt-map: standard PCI properties
|
|
+ to define the mapping of the PCIe interface to interrupt
|
|
+ numbers.
|
|
+- clocks: from common clock binding: handle to pci clock.
|
|
+
|
|
+Optional properties:
|
|
+- status: Either "ok" or "disabled".
|
|
+
|
|
+Example:
|
|
+
|
|
+SoC specific DT Entry:
|
|
+ pcie0: pcie@1f2b0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2b0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xe0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x00 0x00000000 0xe0 0x00000000 0x00 0x00010000 /* io */
|
|
+ 0x02000000 0x00 0x10000000 0xe0 0x10000000 0x00 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xc2 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xc3 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xc4 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xc5 0x1>;
|
|
+ clocks = <&pcie0clk 0>;
|
|
+ };
|
|
+
|
|
+Board specific DT Entry:
|
|
+ &pcie0 {
|
|
+ status = "ok";
|
|
+ };
|
|
diff --git a/MAINTAINERS b/MAINTAINERS
|
|
index 7e2eb4c..7bf051dd 100644
|
|
--- a/MAINTAINERS
|
|
+++ b/MAINTAINERS
|
|
@@ -719,6 +719,14 @@ S: Maintained
|
|
F: drivers/net/appletalk/
|
|
F: net/appletalk/
|
|
|
|
+APPLIED MICRO (APM) X-GENE SOC ETHERNET DRIVER
|
|
+M: Iyappan Subramanian <isubramanian@apm.com>
|
|
+M: Keyur Chudgar <kchudgar@apm.com>
|
|
+M: Ravi Patel <rapatel@apm.com>
|
|
+S: Supported
|
|
+F: drivers/net/ethernet/apm/xgene/
|
|
+F: Documentation/devicetree/bindings/net/apm-xgene-enet.txt
|
|
+
|
|
APTINA CAMERA SENSOR PLL
|
|
M: Laurent Pinchart <Laurent.pinchart@ideasonboard.com>
|
|
L: linux-media@vger.kernel.org
|
|
@@ -6880,6 +6888,13 @@ S: Maintained
|
|
F: Documentation/devicetree/bindings/pci/host-generic-pci.txt
|
|
F: drivers/pci/host/pci-host-generic.c
|
|
|
|
+PCI DRIVER FOR APPLIEDMICRO XGENE
|
|
+M: Tanmay Inamdar <tinamdar@apm.com>
|
|
+L: linux-pci@vger.kernel.org
|
|
+L: linux-arm-kernel@lists.infradead.org
|
|
+S: Maintained
|
|
+F: drivers/pci/host/pci-xgene.c
|
|
+
|
|
PCIE DRIVER FOR ST SPEAR13XX
|
|
M: Mohit Kumar <mohit.kumar@st.com>
|
|
L: linux-pci@vger.kernel.org
|
|
diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig
|
|
index fd4e81a..e8559bb 100644
|
|
--- a/arch/arm64/Kconfig
|
|
+++ b/arch/arm64/Kconfig
|
|
@@ -81,7 +81,7 @@ config MMU
|
|
def_bool y
|
|
|
|
config NO_IOPORT_MAP
|
|
- def_bool y
|
|
+ def_bool y if !PCI
|
|
|
|
config STACKTRACE_SUPPORT
|
|
def_bool y
|
|
@@ -156,6 +156,23 @@ menu "Bus support"
|
|
config ARM_AMBA
|
|
bool
|
|
|
|
+config PCI
|
|
+ bool "PCI support"
|
|
+ help
|
|
+ This feature enables support for PCIe bus system. If you say Y
|
|
+ here, the kernel will include drivers and infrastructure code
|
|
+ to support PCIe bus devices.
|
|
+
|
|
+config PCI_DOMAINS
|
|
+ def_bool PCI
|
|
+
|
|
+config PCI_SYSCALL
|
|
+ def_bool PCI
|
|
+
|
|
+source "drivers/pci/Kconfig"
|
|
+source "drivers/pci/pcie/Kconfig"
|
|
+source "drivers/pci/hotplug/Kconfig"
|
|
+
|
|
endmenu
|
|
|
|
menu "Kernel Features"
|
|
diff --git a/arch/arm64/boot/dts/apm-mustang.dts b/arch/arm64/boot/dts/apm-mustang.dts
|
|
index 6541962..0cb67fc 100644
|
|
--- a/arch/arm64/boot/dts/apm-mustang.dts
|
|
+++ b/arch/arm64/boot/dts/apm-mustang.dts
|
|
@@ -28,3 +28,15 @@
|
|
&serial0 {
|
|
status = "ok";
|
|
};
|
|
+
|
|
+&menet {
|
|
+ status = "ok";
|
|
+};
|
|
+
|
|
+&pcie0clk {
|
|
+ status = "ok";
|
|
+};
|
|
+
|
|
+&pcie0 {
|
|
+ status = "ok";
|
|
+};
|
|
diff --git a/arch/arm64/boot/dts/apm-storm.dtsi b/arch/arm64/boot/dts/apm-storm.dtsi
|
|
index 40aa96c..fb2ee54 100644
|
|
--- a/arch/arm64/boot/dts/apm-storm.dtsi
|
|
+++ b/arch/arm64/boot/dts/apm-storm.dtsi
|
|
@@ -24,56 +24,56 @@
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x000>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@001 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x001>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@100 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x100>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@101 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x101>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@200 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x200>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@201 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x201>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@300 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x300>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
cpu@301 {
|
|
device_type = "cpu";
|
|
compatible = "apm,potenza", "arm,armv8";
|
|
reg = <0x0 0x301>;
|
|
enable-method = "spin-table";
|
|
- cpu-release-addr = <0x1 0x0000fff8>;
|
|
+ cpu-release-addr = <0x40 0x0000f000>;
|
|
};
|
|
};
|
|
|
|
@@ -97,6 +97,11 @@
|
|
clock-frequency = <50000000>;
|
|
};
|
|
|
|
+ pmu {
|
|
+ compatible = "arm,armv8-pmuv3";
|
|
+ interrupts = <1 12 0xff04>;
|
|
+ };
|
|
+
|
|
soc {
|
|
compatible = "simple-bus";
|
|
#address-cells = <2>;
|
|
@@ -167,14 +172,13 @@
|
|
clock-output-names = "ethclk";
|
|
};
|
|
|
|
- eth8clk: eth8clk {
|
|
+ menetclk: menetclk {
|
|
compatible = "apm,xgene-device-clock";
|
|
#clock-cells = <1>;
|
|
clocks = <ðclk 0>;
|
|
- clock-names = "eth8clk";
|
|
reg = <0x0 0x1702C000 0x0 0x1000>;
|
|
reg-names = "csr-reg";
|
|
- clock-output-names = "eth8clk";
|
|
+ clock-output-names = "menetclk";
|
|
};
|
|
|
|
sataphy1clk: sataphy1clk@1f21c000 {
|
|
@@ -270,6 +274,161 @@
|
|
enable-mask = <0x2>;
|
|
clock-output-names = "rtcclk";
|
|
};
|
|
+
|
|
+ pcie0clk: pcie0clk@1f2bc000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f2bc000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie0clk";
|
|
+ };
|
|
+
|
|
+ pcie1clk: pcie1clk@1f2cc000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f2cc000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie1clk";
|
|
+ };
|
|
+
|
|
+ pcie2clk: pcie2clk@1f2dc000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f2dc000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie2clk";
|
|
+ };
|
|
+
|
|
+ pcie3clk: pcie3clk@1f50c000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f50c000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie3clk";
|
|
+ };
|
|
+
|
|
+ pcie4clk: pcie4clk@1f51c000 {
|
|
+ status = "disabled";
|
|
+ compatible = "apm,xgene-device-clock";
|
|
+ #clock-cells = <1>;
|
|
+ clocks = <&socplldiv2 0>;
|
|
+ reg = <0x0 0x1f51c000 0x0 0x1000>;
|
|
+ reg-names = "csr-reg";
|
|
+ clock-output-names = "pcie4clk";
|
|
+ };
|
|
+ };
|
|
+
|
|
+ pcie0: pcie@1f2b0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2b0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xe0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x00 0x00000000 0xe0 0x00000000 0x00 0x00010000 /* io */
|
|
+ 0x02000000 0x00 0x10000000 0xe0 0x10000000 0x00 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xc2 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xc3 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xc4 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xc5 0x1>;
|
|
+ clocks = <&pcie0clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie1: pcie@1f2c0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2c0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xd0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0xd0 0x00000000 0x00 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0xd0 0x10000000 0x00 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xc8 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xc9 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xca 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xcb 0x1>;
|
|
+ clocks = <&pcie1clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie2: pcie@1f2d0000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f2d0000 0x0 0x00010000 /* Controller registers */
|
|
+ 0x90 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0x90 0x00000000 0x0 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0x90 0x10000000 0x0 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xce 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xcf 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xd0 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xd1 0x1>;
|
|
+ clocks = <&pcie2clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie3: pcie@1f500000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f500000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xa0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0xa0 0x00000000 0x0 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0xa0 0x10000000 0x0 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xd4 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xd5 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xd6 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xd7 0x1>;
|
|
+ clocks = <&pcie3clk 0>;
|
|
+ };
|
|
+
|
|
+ pcie4: pcie@1f510000 {
|
|
+ status = "disabled";
|
|
+ device_type = "pci";
|
|
+ compatible = "apm,xgene-storm-pcie", "apm,xgene-pcie";
|
|
+ #interrupt-cells = <1>;
|
|
+ #size-cells = <2>;
|
|
+ #address-cells = <3>;
|
|
+ reg = < 0x00 0x1f510000 0x0 0x00010000 /* Controller registers */
|
|
+ 0xc0 0xd0000000 0x0 0x00200000>; /* PCI config space */
|
|
+ reg-names = "csr", "cfg";
|
|
+ ranges = <0x01000000 0x0 0x00000000 0xc0 0x00000000 0x0 0x00010000 /* io */
|
|
+ 0x02000000 0x0 0x10000000 0xc0 0x10000000 0x0 0x80000000>; /* mem */
|
|
+ dma-ranges = <0x42000000 0x40 0x00000000 0x40 0x00000000 0x40 0x00000000>;
|
|
+ interrupt-map-mask = <0x0 0x0 0x0 0x7>;
|
|
+ interrupt-map = <0x0 0x0 0x0 0x1 &gic 0x0 0xda 0x1
|
|
+ 0x0 0x0 0x0 0x2 &gic 0x0 0xdb 0x1
|
|
+ 0x0 0x0 0x0 0x3 &gic 0x0 0xdc 0x1
|
|
+ 0x0 0x0 0x0 0x4 &gic 0x0 0xdd 0x1>;
|
|
+ clocks = <&pcie4clk 0>;
|
|
};
|
|
|
|
serial0: serial@1c020000 {
|
|
@@ -278,7 +437,7 @@
|
|
compatible = "ns16550a";
|
|
reg = <0 0x1c020000 0x0 0x1000>;
|
|
reg-shift = <2>;
|
|
- clock-frequency = <10000000>; /* Updated by bootloader */
|
|
+ clock-frequency = <50000000>; /* Updated by bootloader */
|
|
interrupt-parent = <&gic>;
|
|
interrupts = <0x0 0x4c 0x4>;
|
|
};
|
|
@@ -397,5 +556,30 @@
|
|
#clock-cells = <1>;
|
|
clocks = <&rtcclk 0>;
|
|
};
|
|
+
|
|
+ menet: ethernet@17020000 {
|
|
+ compatible = "apm,xgene-enet";
|
|
+ status = "disabled";
|
|
+ reg = <0x0 0x17020000 0x0 0xd100>,
|
|
+ <0x0 0X17030000 0x0 0X400>,
|
|
+ <0x0 0X10000000 0x0 0X200>;
|
|
+ reg-names = "enet_csr", "ring_csr", "ring_cmd";
|
|
+ interrupts = <0x0 0x3c 0x4>;
|
|
+ dma-coherent;
|
|
+ clocks = <&menetclk 0>;
|
|
+ local-mac-address = [00 00 00 00 00 00];
|
|
+ phy-connection-type = "rgmii";
|
|
+ phy-handle = <&menetphy>;
|
|
+ mdio {
|
|
+ compatible = "apm,xgene-mdio";
|
|
+ #address-cells = <1>;
|
|
+ #size-cells = <0>;
|
|
+ menetphy: menetphy@3 {
|
|
+ compatible = "ethernet-phy-id001c.c915";
|
|
+ reg = <0x3>;
|
|
+ };
|
|
+
|
|
+ };
|
|
+ };
|
|
};
|
|
};
|
|
diff --git a/arch/arm64/include/asm/Kbuild b/arch/arm64/include/asm/Kbuild
|
|
index 0b3fcf8..07cb417 100644
|
|
--- a/arch/arm64/include/asm/Kbuild
|
|
+++ b/arch/arm64/include/asm/Kbuild
|
|
@@ -29,6 +29,7 @@ generic-y += mman.h
|
|
generic-y += msgbuf.h
|
|
generic-y += mutex.h
|
|
generic-y += pci.h
|
|
+generic-y += pci-bridge.h
|
|
generic-y += poll.h
|
|
generic-y += preempt.h
|
|
generic-y += resource.h
|
|
diff --git a/arch/arm64/include/asm/elf.h b/arch/arm64/include/asm/elf.h
|
|
index 01d3aab..8186df6 100644
|
|
--- a/arch/arm64/include/asm/elf.h
|
|
+++ b/arch/arm64/include/asm/elf.h
|
|
@@ -114,7 +114,8 @@ typedef struct user_fpsimd_state elf_fpregset_t;
|
|
*/
|
|
#define elf_check_arch(x) ((x)->e_machine == EM_AARCH64)
|
|
|
|
-#define elf_read_implies_exec(ex,stk) (stk != EXSTACK_DISABLE_X)
|
|
+#define elf_read_implies_exec(ex,stk) (test_thread_flag(TIF_32BIT) \
|
|
+ ? (stk == EXSTACK_ENABLE_X) : 0)
|
|
|
|
#define CORE_DUMP_USE_REGSET
|
|
#define ELF_EXEC_PAGESIZE PAGE_SIZE
|
|
diff --git a/arch/arm64/include/asm/io.h b/arch/arm64/include/asm/io.h
|
|
index e0ecdcf..dc34039 100644
|
|
--- a/arch/arm64/include/asm/io.h
|
|
+++ b/arch/arm64/include/asm/io.h
|
|
@@ -121,7 +121,8 @@ static inline u64 __raw_readq(const volatile void __iomem *addr)
|
|
/*
|
|
* I/O port access primitives.
|
|
*/
|
|
-#define IO_SPACE_LIMIT 0xffff
|
|
+#define arch_has_dev_port() (1)
|
|
+#define IO_SPACE_LIMIT 0x1ffffff
|
|
#define PCI_IOBASE ((void __iomem *)(MODULES_VADDR - SZ_32M))
|
|
|
|
static inline u8 inb(unsigned long addr)
|
|
diff --git a/arch/arm64/include/asm/kvm_mmu.h b/arch/arm64/include/asm/kvm_mmu.h
|
|
index 8e138c7..f8c9385 100644
|
|
--- a/arch/arm64/include/asm/kvm_mmu.h
|
|
+++ b/arch/arm64/include/asm/kvm_mmu.h
|
|
@@ -122,8 +122,16 @@ static inline void kvm_set_s2pmd_writable(pmd_t *pmd)
|
|
}
|
|
|
|
#define kvm_pgd_addr_end(addr, end) pgd_addr_end(addr, end)
|
|
-#define kvm_pud_addr_end(addr, end) pud_addr_end(addr, end)
|
|
-#define kvm_pmd_addr_end(addr, end) pmd_addr_end(addr, end)
|
|
+
|
|
+#define kvm_pud_addr_end(addr, end) \
|
|
+({ unsigned long __boundary = ((addr) + PUD_SIZE) & PUD_MASK; \
|
|
+ (__boundary - 1 < (end) - 1)? __boundary: (end); \
|
|
+})
|
|
+
|
|
+#define kvm_pmd_addr_end(addr, end) \
|
|
+({ unsigned long __boundary = ((addr) + PMD_SIZE) & PMD_MASK; \
|
|
+ (__boundary - 1 < (end) - 1)? __boundary: (end); \
|
|
+})
|
|
|
|
static inline bool kvm_page_empty(void *ptr)
|
|
{
|
|
diff --git a/arch/arm64/include/asm/pci.h b/arch/arm64/include/asm/pci.h
|
|
new file mode 100644
|
|
index 0000000..3f7856e
|
|
--- /dev/null
|
|
+++ b/arch/arm64/include/asm/pci.h
|
|
@@ -0,0 +1,49 @@
|
|
+#ifndef __ASM_PCI_H
|
|
+#define __ASM_PCI_H
|
|
+#ifdef __KERNEL__
|
|
+
|
|
+#include <linux/types.h>
|
|
+#include <linux/slab.h>
|
|
+#include <linux/dma-mapping.h>
|
|
+
|
|
+#include <asm/io.h>
|
|
+#include <asm-generic/pci-bridge.h>
|
|
+#include <asm-generic/pci-dma-compat.h>
|
|
+
|
|
+#define PCIBIOS_MIN_IO 0x1000
|
|
+#define PCIBIOS_MIN_MEM 0
|
|
+
|
|
+struct pci_host_bridge *find_pci_host_bridge(struct pci_bus *bus);
|
|
+
|
|
+/*
|
|
+ * Set to 1 if the kernel should re-assign all PCI bus numbers
|
|
+ */
|
|
+#define pcibios_assign_all_busses() \
|
|
+ (pci_has_flag(PCI_REASSIGN_ALL_BUS))
|
|
+
|
|
+/*
|
|
+ * PCI address space differs from physical memory address space
|
|
+ */
|
|
+#define PCI_DMA_BUS_IS_PHYS (0)
|
|
+
|
|
+extern int isa_dma_bridge_buggy;
|
|
+
|
|
+#ifdef CONFIG_PCI
|
|
+static inline int pci_domain_nr(struct pci_bus *bus)
|
|
+{
|
|
+ struct pci_host_bridge *bridge = find_pci_host_bridge(bus);
|
|
+
|
|
+ if (bridge)
|
|
+ return bridge->domain_nr;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static inline int pci_proc_domain(struct pci_bus *bus)
|
|
+{
|
|
+ return 1;
|
|
+}
|
|
+#endif /* CONFIG_PCI */
|
|
+
|
|
+#endif /* __KERNEL__ */
|
|
+#endif /* __ASM_PCI_H */
|
|
diff --git a/arch/arm64/kernel/Makefile b/arch/arm64/kernel/Makefile
|
|
index df7ef87..1ed5a06 100644
|
|
--- a/arch/arm64/kernel/Makefile
|
|
+++ b/arch/arm64/kernel/Makefile
|
|
@@ -29,6 +29,7 @@ arm64-obj-$(CONFIG_ARM64_CPU_SUSPEND) += sleep.o suspend.o
|
|
arm64-obj-$(CONFIG_JUMP_LABEL) += jump_label.o
|
|
arm64-obj-$(CONFIG_KGDB) += kgdb.o
|
|
arm64-obj-$(CONFIG_EFI) += efi.o efi-stub.o efi-entry.o
|
|
+arm64-obj-$(CONFIG_PCI) += pci.o
|
|
|
|
obj-y += $(arm64-obj-y) vdso/
|
|
obj-m += $(arm64-obj-m)
|
|
diff --git a/arch/arm64/kernel/efi-stub.c b/arch/arm64/kernel/efi-stub.c
|
|
index 1317fef..d27dd98 100644
|
|
--- a/arch/arm64/kernel/efi-stub.c
|
|
+++ b/arch/arm64/kernel/efi-stub.c
|
|
@@ -28,20 +28,16 @@ efi_status_t handle_kernel_image(efi_system_table_t *sys_table,
|
|
kernel_size = _edata - _text;
|
|
if (*image_addr != (dram_base + TEXT_OFFSET)) {
|
|
kernel_memsize = kernel_size + (_end - _edata);
|
|
- status = efi_relocate_kernel(sys_table, image_addr,
|
|
- kernel_size, kernel_memsize,
|
|
- dram_base + TEXT_OFFSET,
|
|
- PAGE_SIZE);
|
|
+ status = efi_low_alloc(sys_table, kernel_memsize + TEXT_OFFSET,
|
|
+ SZ_2M, reserve_addr);
|
|
if (status != EFI_SUCCESS) {
|
|
pr_efi_err(sys_table, "Failed to relocate kernel\n");
|
|
return status;
|
|
}
|
|
- if (*image_addr != (dram_base + TEXT_OFFSET)) {
|
|
- pr_efi_err(sys_table, "Failed to alloc kernel memory\n");
|
|
- efi_free(sys_table, kernel_memsize, *image_addr);
|
|
- return EFI_LOAD_ERROR;
|
|
- }
|
|
- *image_size = kernel_memsize;
|
|
+ memcpy((void *)*reserve_addr + TEXT_OFFSET, (void *)*image_addr,
|
|
+ kernel_size);
|
|
+ *image_addr = *reserve_addr + TEXT_OFFSET;
|
|
+ *reserve_size = kernel_memsize + TEXT_OFFSET;
|
|
}
|
|
|
|
|
|
diff --git a/arch/arm64/kernel/efi.c b/arch/arm64/kernel/efi.c
|
|
index e72f310..72ee260 100644
|
|
--- a/arch/arm64/kernel/efi.c
|
|
+++ b/arch/arm64/kernel/efi.c
|
|
@@ -475,3 +475,14 @@ err_unmap:
|
|
return -1;
|
|
}
|
|
early_initcall(arm64_enter_virtual_mode);
|
|
+
|
|
+/*
|
|
+ * If nothing else is handling pm_power_off, use EFI
|
|
+ *
|
|
+ * This is called from a late_initcall after other mechanisms
|
|
+ * have had a chance to register a handler.
|
|
+ */
|
|
+bool efi_poweroff_required(void)
|
|
+{
|
|
+ return pm_power_off == NULL;
|
|
+}
|
|
diff --git a/arch/arm64/kernel/head.S b/arch/arm64/kernel/head.S
|
|
index 144f105..b6ca95a 100644
|
|
--- a/arch/arm64/kernel/head.S
|
|
+++ b/arch/arm64/kernel/head.S
|
|
@@ -151,7 +151,7 @@ optional_header:
|
|
.short 0x20b // PE32+ format
|
|
.byte 0x02 // MajorLinkerVersion
|
|
.byte 0x14 // MinorLinkerVersion
|
|
- .long _edata - stext // SizeOfCode
|
|
+ .long _end - stext // SizeOfCode
|
|
.long 0 // SizeOfInitializedData
|
|
.long 0 // SizeOfUninitializedData
|
|
.long efi_stub_entry - efi_head // AddressOfEntryPoint
|
|
@@ -169,7 +169,7 @@ extra_header_fields:
|
|
.short 0 // MinorSubsystemVersion
|
|
.long 0 // Win32VersionValue
|
|
|
|
- .long _edata - efi_head // SizeOfImage
|
|
+ .long _end - efi_head // SizeOfImage
|
|
|
|
// Everything before the kernel image is considered part of the header
|
|
.long stext - efi_head // SizeOfHeaders
|
|
@@ -216,7 +216,7 @@ section_table:
|
|
.byte 0
|
|
.byte 0
|
|
.byte 0 // end of 0 padding of section name
|
|
- .long _edata - stext // VirtualSize
|
|
+ .long _end - stext // VirtualSize
|
|
.long stext - efi_head // VirtualAddress
|
|
.long _edata - stext // SizeOfRawData
|
|
.long stext - efi_head // PointerToRawData
|
|
diff --git a/arch/arm64/kernel/pci.c b/arch/arm64/kernel/pci.c
|
|
new file mode 100644
|
|
index 0000000..955d6d1
|
|
--- /dev/null
|
|
+++ b/arch/arm64/kernel/pci.c
|
|
@@ -0,0 +1,38 @@
|
|
+/*
|
|
+ * Code borrowed from powerpc/kernel/pci-common.c
|
|
+ *
|
|
+ * Copyright (C) 2003 Anton Blanchard <anton@au.ibm.com>, IBM
|
|
+ * Copyright (C) 2014 ARM Ltd.
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or
|
|
+ * modify it under the terms of the GNU General Public License
|
|
+ * version 2 as published by the Free Software Foundation.
|
|
+ *
|
|
+ */
|
|
+
|
|
+#include <linux/init.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/kernel.h>
|
|
+#include <linux/mm.h>
|
|
+#include <linux/of_pci.h>
|
|
+#include <linux/of_platform.h>
|
|
+#include <linux/slab.h>
|
|
+
|
|
+#include <asm/pci-bridge.h>
|
|
+
|
|
+/*
|
|
+ * Called after each bus is probed, but before its children are examined
|
|
+ */
|
|
+void pcibios_fixup_bus(struct pci_bus *bus)
|
|
+{
|
|
+ /* nothing to do, expected to be removed in the future */
|
|
+}
|
|
+
|
|
+/*
|
|
+ * We don't have to worry about legacy ISA devices, so nothing to do here
|
|
+ */
|
|
+resource_size_t pcibios_align_resource(void *data, const struct resource *res,
|
|
+ resource_size_t size, resource_size_t align)
|
|
+{
|
|
+ return res->start;
|
|
+}
|
|
diff --git a/arch/arm64/kernel/process.c b/arch/arm64/kernel/process.c
|
|
index 1309d64..8ff5208 100644
|
|
--- a/arch/arm64/kernel/process.c
|
|
+++ b/arch/arm64/kernel/process.c
|
|
@@ -43,6 +43,7 @@
|
|
#include <linux/hw_breakpoint.h>
|
|
#include <linux/personality.h>
|
|
#include <linux/notifier.h>
|
|
+#include <linux/efi.h>
|
|
|
|
#include <asm/compat.h>
|
|
#include <asm/cacheflush.h>
|
|
@@ -182,6 +183,11 @@ void machine_restart(char *cmd)
|
|
arm_pm_restart(reboot_mode, cmd);
|
|
|
|
/*
|
|
+ * If all else fails, try EFI
|
|
+ */
|
|
+ efi_reboot(reboot_mode, cmd);
|
|
+
|
|
+ /*
|
|
* Whoops - the architecture was unable to reboot.
|
|
*/
|
|
printk("Reboot failed -- System halted\n");
|
|
diff --git a/arch/arm64/kernel/smp_spin_table.c b/arch/arm64/kernel/smp_spin_table.c
|
|
index 0347d38..70181c1 100644
|
|
--- a/arch/arm64/kernel/smp_spin_table.c
|
|
+++ b/arch/arm64/kernel/smp_spin_table.c
|
|
@@ -20,6 +20,7 @@
|
|
#include <linux/init.h>
|
|
#include <linux/of.h>
|
|
#include <linux/smp.h>
|
|
+#include <linux/types.h>
|
|
|
|
#include <asm/cacheflush.h>
|
|
#include <asm/cpu_ops.h>
|
|
@@ -65,12 +66,21 @@ static int smp_spin_table_cpu_init(struct device_node *dn, unsigned int cpu)
|
|
|
|
static int smp_spin_table_cpu_prepare(unsigned int cpu)
|
|
{
|
|
- void **release_addr;
|
|
+ __le64 __iomem *release_addr;
|
|
|
|
if (!cpu_release_addr[cpu])
|
|
return -ENODEV;
|
|
|
|
- release_addr = __va(cpu_release_addr[cpu]);
|
|
+ /*
|
|
+ * The cpu-release-addr may or may not be inside the linear mapping.
|
|
+ * As ioremap_cache will either give us a new mapping or reuse the
|
|
+ * existing linear mapping, we can use it to cover both cases. In
|
|
+ * either case the memory will be MT_NORMAL.
|
|
+ */
|
|
+ release_addr = ioremap_cache(cpu_release_addr[cpu],
|
|
+ sizeof(*release_addr));
|
|
+ if (!release_addr)
|
|
+ return -ENOMEM;
|
|
|
|
/*
|
|
* We write the release address as LE regardless of the native
|
|
@@ -79,15 +89,16 @@ static int smp_spin_table_cpu_prepare(unsigned int cpu)
|
|
* boot-loader's endianess before jumping. This is mandated by
|
|
* the boot protocol.
|
|
*/
|
|
- release_addr[0] = (void *) cpu_to_le64(__pa(secondary_holding_pen));
|
|
-
|
|
- __flush_dcache_area(release_addr, sizeof(release_addr[0]));
|
|
+ writeq_relaxed(__pa(secondary_holding_pen), release_addr);
|
|
+ __flush_dcache_area(release_addr, sizeof(*release_addr));
|
|
|
|
/*
|
|
* Send an event to wake up the secondary CPU.
|
|
*/
|
|
sev();
|
|
|
|
+ iounmap(release_addr);
|
|
+
|
|
return 0;
|
|
}
|
|
|
|
diff --git a/drivers/ata/ahci_xgene.c b/drivers/ata/ahci_xgene.c
|
|
index bc28111..00e5a0c 100644
|
|
--- a/drivers/ata/ahci_xgene.c
|
|
+++ b/drivers/ata/ahci_xgene.c
|
|
@@ -134,7 +134,8 @@ static unsigned int xgene_ahci_qc_issue(struct ata_queued_cmd *qc)
|
|
struct xgene_ahci_context *ctx = hpriv->plat_data;
|
|
int rc = 0;
|
|
|
|
- if (unlikely(ctx->last_cmd[ap->port_no] == ATA_CMD_ID_ATA))
|
|
+ if (unlikely(ctx->last_cmd[ap->port_no] == ATA_CMD_ID_ATA ||
|
|
+ ctx->last_cmd[ap->port_no] == ATA_CMD_SMART))
|
|
xgene_ahci_restart_engine(ap);
|
|
|
|
rc = ahci_qc_issue(qc);
|
|
diff --git a/drivers/irqchip/irq-gic.c b/drivers/irqchip/irq-gic.c
|
|
index 4b959e6..c36c7ab55 100644
|
|
--- a/drivers/irqchip/irq-gic.c
|
|
+++ b/drivers/irqchip/irq-gic.c
|
|
@@ -381,6 +381,7 @@ static void gic_cpu_init(struct gic_chip_data *gic)
|
|
void __iomem *dist_base = gic_data_dist_base(gic);
|
|
void __iomem *base = gic_data_cpu_base(gic);
|
|
unsigned int cpu_mask, cpu = smp_processor_id();
|
|
+ unsigned int ctrl_mask;
|
|
int i;
|
|
|
|
/*
|
|
@@ -401,13 +402,29 @@ static void gic_cpu_init(struct gic_chip_data *gic)
|
|
gic_cpu_config(dist_base, NULL);
|
|
|
|
writel_relaxed(0xf0, base + GIC_CPU_PRIMASK);
|
|
- writel_relaxed(1, base + GIC_CPU_CTRL);
|
|
+
|
|
+ ctrl_mask = readl(base + GIC_CPU_CTRL);
|
|
+
|
|
+ /* Mask out the gic v2 bypass bits */
|
|
+ ctrl_mask &= 0x1e0;
|
|
+
|
|
+ /* Enable group 0 */
|
|
+ ctrl_mask |= 0x1;
|
|
+ writel_relaxed(ctrl_mask, base + GIC_CPU_CTRL);
|
|
}
|
|
|
|
void gic_cpu_if_down(void)
|
|
{
|
|
+ unsigned int ctrl_mask;
|
|
void __iomem *cpu_base = gic_data_cpu_base(&gic_data[0]);
|
|
- writel_relaxed(0, cpu_base + GIC_CPU_CTRL);
|
|
+
|
|
+ ctrl_mask = readl(cpu_base + GIC_CPU_CTRL);
|
|
+ /*
|
|
+ * Disable grp enable bit, leave the bypass bits alone as changing
|
|
+ * them could leave the system unstable
|
|
+ */
|
|
+ ctrl_mask &= 0x1e0;
|
|
+ writel_relaxed(ctrl_mask, cpu_base + GIC_CPU_CTRL);
|
|
}
|
|
|
|
#ifdef CONFIG_CPU_PM
|
|
@@ -518,6 +535,7 @@ static void gic_cpu_restore(unsigned int gic_nr)
|
|
{
|
|
int i;
|
|
u32 *ptr;
|
|
+ unsigned int ctrl_mask;
|
|
void __iomem *dist_base;
|
|
void __iomem *cpu_base;
|
|
|
|
@@ -542,7 +560,15 @@ static void gic_cpu_restore(unsigned int gic_nr)
|
|
writel_relaxed(0xa0a0a0a0, dist_base + GIC_DIST_PRI + i * 4);
|
|
|
|
writel_relaxed(0xf0, cpu_base + GIC_CPU_PRIMASK);
|
|
- writel_relaxed(1, cpu_base + GIC_CPU_CTRL);
|
|
+
|
|
+ ctrl_mask = readl(cpu_base + GIC_CPU_CTRL);
|
|
+
|
|
+ /* Mask out the gic v2 bypass bits */
|
|
+ ctrl_mask &= 0x1e0;
|
|
+
|
|
+ /* Enable group 0 */
|
|
+ ctrl_mask |= 0x1;
|
|
+ writel_relaxed(ctrl_mask, cpu_base + GIC_CPU_CTRL);
|
|
}
|
|
|
|
static int gic_notifier(struct notifier_block *self, unsigned long cmd, void *v)
|
|
diff --git a/drivers/net/ethernet/Kconfig b/drivers/net/ethernet/Kconfig
|
|
index edb7186..dc7406c 100644
|
|
--- a/drivers/net/ethernet/Kconfig
|
|
+++ b/drivers/net/ethernet/Kconfig
|
|
@@ -24,6 +24,7 @@ source "drivers/net/ethernet/allwinner/Kconfig"
|
|
source "drivers/net/ethernet/alteon/Kconfig"
|
|
source "drivers/net/ethernet/altera/Kconfig"
|
|
source "drivers/net/ethernet/amd/Kconfig"
|
|
+source "drivers/net/ethernet/apm/Kconfig"
|
|
source "drivers/net/ethernet/apple/Kconfig"
|
|
source "drivers/net/ethernet/arc/Kconfig"
|
|
source "drivers/net/ethernet/atheros/Kconfig"
|
|
diff --git a/drivers/net/ethernet/Makefile b/drivers/net/ethernet/Makefile
|
|
index 58de333..224a018 100644
|
|
--- a/drivers/net/ethernet/Makefile
|
|
+++ b/drivers/net/ethernet/Makefile
|
|
@@ -10,6 +10,7 @@ obj-$(CONFIG_NET_VENDOR_ALLWINNER) += allwinner/
|
|
obj-$(CONFIG_NET_VENDOR_ALTEON) += alteon/
|
|
obj-$(CONFIG_ALTERA_TSE) += altera/
|
|
obj-$(CONFIG_NET_VENDOR_AMD) += amd/
|
|
+obj-$(CONFIG_NET_XGENE) += apm/
|
|
obj-$(CONFIG_NET_VENDOR_APPLE) += apple/
|
|
obj-$(CONFIG_NET_VENDOR_ARC) += arc/
|
|
obj-$(CONFIG_NET_VENDOR_ATHEROS) += atheros/
|
|
diff --git a/drivers/net/ethernet/apm/Kconfig b/drivers/net/ethernet/apm/Kconfig
|
|
new file mode 100644
|
|
index 0000000..ec63d70
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/Kconfig
|
|
@@ -0,0 +1 @@
|
|
+source "drivers/net/ethernet/apm/xgene/Kconfig"
|
|
diff --git a/drivers/net/ethernet/apm/Makefile b/drivers/net/ethernet/apm/Makefile
|
|
new file mode 100644
|
|
index 0000000..65ce32a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/Makefile
|
|
@@ -0,0 +1,5 @@
|
|
+#
|
|
+# Makefile for APM X-GENE Ethernet driver.
|
|
+#
|
|
+
|
|
+obj-$(CONFIG_NET_XGENE) += xgene/
|
|
diff --git a/drivers/net/ethernet/apm/xgene/Kconfig b/drivers/net/ethernet/apm/xgene/Kconfig
|
|
new file mode 100644
|
|
index 0000000..616dff6
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/Kconfig
|
|
@@ -0,0 +1,9 @@
|
|
+config NET_XGENE
|
|
+ tristate "APM X-Gene SoC Ethernet Driver"
|
|
+ select PHYLIB
|
|
+ help
|
|
+ This is the Ethernet driver for the on-chip ethernet interface on the
|
|
+ APM X-Gene SoC.
|
|
+
|
|
+ To compile this driver as a module, choose M here. This module will
|
|
+ be called xgene_enet.
|
|
diff --git a/drivers/net/ethernet/apm/xgene/Makefile b/drivers/net/ethernet/apm/xgene/Makefile
|
|
new file mode 100644
|
|
index 0000000..c643e8a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/Makefile
|
|
@@ -0,0 +1,6 @@
|
|
+#
|
|
+# Makefile for APM X-Gene Ethernet Driver.
|
|
+#
|
|
+
|
|
+xgene-enet-objs := xgene_enet_hw.o xgene_enet_main.o xgene_enet_ethtool.o
|
|
+obj-$(CONFIG_NET_XGENE) += xgene-enet.o
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_ethtool.c b/drivers/net/ethernet/apm/xgene/xgene_enet_ethtool.c
|
|
new file mode 100644
|
|
index 0000000..63f2aa5
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_ethtool.c
|
|
@@ -0,0 +1,125 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include <linux/ethtool.h>
|
|
+#include "xgene_enet_main.h"
|
|
+
|
|
+struct xgene_gstrings_stats {
|
|
+ char name[ETH_GSTRING_LEN];
|
|
+ int offset;
|
|
+};
|
|
+
|
|
+#define XGENE_STAT(m) { #m, offsetof(struct xgene_enet_pdata, stats.m) }
|
|
+
|
|
+static const struct xgene_gstrings_stats gstrings_stats[] = {
|
|
+ XGENE_STAT(rx_packets),
|
|
+ XGENE_STAT(tx_packets),
|
|
+ XGENE_STAT(rx_bytes),
|
|
+ XGENE_STAT(tx_bytes),
|
|
+ XGENE_STAT(rx_errors),
|
|
+ XGENE_STAT(tx_errors),
|
|
+ XGENE_STAT(rx_length_errors),
|
|
+ XGENE_STAT(rx_crc_errors),
|
|
+ XGENE_STAT(rx_frame_errors),
|
|
+ XGENE_STAT(rx_fifo_errors)
|
|
+};
|
|
+
|
|
+#define XGENE_STATS_LEN ARRAY_SIZE(gstrings_stats)
|
|
+
|
|
+static void xgene_get_drvinfo(struct net_device *ndev,
|
|
+ struct ethtool_drvinfo *info)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct platform_device *pdev = pdata->pdev;
|
|
+
|
|
+ strcpy(info->driver, "xgene_enet");
|
|
+ strcpy(info->version, XGENE_DRV_VERSION);
|
|
+ snprintf(info->fw_version, ETHTOOL_FWVERS_LEN, "N/A");
|
|
+ sprintf(info->bus_info, "%s", pdev->name);
|
|
+}
|
|
+
|
|
+static int xgene_get_settings(struct net_device *ndev, struct ethtool_cmd *cmd)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct phy_device *phydev = pdata->phy_dev;
|
|
+
|
|
+ if (phydev == NULL)
|
|
+ return -ENODEV;
|
|
+
|
|
+ return phy_ethtool_gset(phydev, cmd);
|
|
+}
|
|
+
|
|
+static int xgene_set_settings(struct net_device *ndev, struct ethtool_cmd *cmd)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct phy_device *phydev = pdata->phy_dev;
|
|
+
|
|
+ if (phydev == NULL)
|
|
+ return -ENODEV;
|
|
+
|
|
+ return phy_ethtool_sset(phydev, cmd);
|
|
+}
|
|
+
|
|
+static void xgene_get_strings(struct net_device *ndev, u32 stringset, u8 *data)
|
|
+{
|
|
+ int i;
|
|
+ u8 *p = data;
|
|
+
|
|
+ if (stringset != ETH_SS_STATS)
|
|
+ return;
|
|
+
|
|
+ for (i = 0; i < XGENE_STATS_LEN; i++) {
|
|
+ memcpy(p, gstrings_stats[i].name, ETH_GSTRING_LEN);
|
|
+ p += ETH_GSTRING_LEN;
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_get_sset_count(struct net_device *ndev, int sset)
|
|
+{
|
|
+ if (sset != ETH_SS_STATS)
|
|
+ return -EINVAL;
|
|
+
|
|
+ return XGENE_STATS_LEN;
|
|
+}
|
|
+
|
|
+static void xgene_get_ethtool_stats(struct net_device *ndev,
|
|
+ struct ethtool_stats *dummy,
|
|
+ u64 *data)
|
|
+{
|
|
+ void *pdata = netdev_priv(ndev);
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < XGENE_STATS_LEN; i++)
|
|
+ *data++ = *(u64 *)(pdata + gstrings_stats[i].offset);
|
|
+}
|
|
+
|
|
+static const struct ethtool_ops xgene_ethtool_ops = {
|
|
+ .get_drvinfo = xgene_get_drvinfo,
|
|
+ .get_settings = xgene_get_settings,
|
|
+ .set_settings = xgene_set_settings,
|
|
+ .get_link = ethtool_op_get_link,
|
|
+ .get_strings = xgene_get_strings,
|
|
+ .get_sset_count = xgene_get_sset_count,
|
|
+ .get_ethtool_stats = xgene_get_ethtool_stats
|
|
+};
|
|
+
|
|
+void xgene_enet_set_ethtool_ops(struct net_device *ndev)
|
|
+{
|
|
+ ndev->ethtool_ops = &xgene_ethtool_ops;
|
|
+}
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_hw.c b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
|
|
new file mode 100644
|
|
index 0000000..e52af60
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
|
|
@@ -0,0 +1,747 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include "xgene_enet_main.h"
|
|
+#include "xgene_enet_hw.h"
|
|
+
|
|
+static void xgene_enet_ring_init(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *ring_cfg = ring->state;
|
|
+ u64 addr = ring->dma;
|
|
+ enum xgene_enet_ring_cfgsize cfgsize = ring->cfgsize;
|
|
+
|
|
+ ring_cfg[4] |= (1 << SELTHRSH_POS) &
|
|
+ CREATE_MASK(SELTHRSH_POS, SELTHRSH_LEN);
|
|
+ ring_cfg[3] |= ACCEPTLERR;
|
|
+ ring_cfg[2] |= QCOHERENT;
|
|
+
|
|
+ addr >>= 8;
|
|
+ ring_cfg[2] |= (addr << RINGADDRL_POS) &
|
|
+ CREATE_MASK_ULL(RINGADDRL_POS, RINGADDRL_LEN);
|
|
+ addr >>= RINGADDRL_LEN;
|
|
+ ring_cfg[3] |= addr & CREATE_MASK_ULL(RINGADDRH_POS, RINGADDRH_LEN);
|
|
+ ring_cfg[3] |= ((u32) cfgsize << RINGSIZE_POS) &
|
|
+ CREATE_MASK(RINGSIZE_POS, RINGSIZE_LEN);
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_set_type(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *ring_cfg = ring->state;
|
|
+ bool is_bufpool;
|
|
+ u32 val;
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+ val = (is_bufpool) ? RING_BUFPOOL : RING_REGULAR;
|
|
+ ring_cfg[4] |= (val << RINGTYPE_POS) &
|
|
+ CREATE_MASK(RINGTYPE_POS, RINGTYPE_LEN);
|
|
+
|
|
+ if (is_bufpool) {
|
|
+ ring_cfg[3] |= (BUFPOOL_MODE << RINGMODE_POS) &
|
|
+ CREATE_MASK(RINGMODE_POS, RINGMODE_LEN);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_set_recombbuf(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *ring_cfg = ring->state;
|
|
+
|
|
+ ring_cfg[3] |= RECOMBBUF;
|
|
+ ring_cfg[3] |= (0xf << RECOMTIMEOUTL_POS) &
|
|
+ CREATE_MASK(RECOMTIMEOUTL_POS, RECOMTIMEOUTL_LEN);
|
|
+ ring_cfg[4] |= 0x7 & CREATE_MASK(RECOMTIMEOUTH_POS, RECOMTIMEOUTH_LEN);
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_wr32(struct xgene_enet_desc_ring *ring,
|
|
+ u32 offset, u32 data)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+
|
|
+ iowrite32(data, pdata->ring_csr_addr + offset);
|
|
+}
|
|
+
|
|
+static void xgene_enet_ring_rd32(struct xgene_enet_desc_ring *ring,
|
|
+ u32 offset, u32 *data)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+
|
|
+ *data = ioread32(pdata->ring_csr_addr + offset);
|
|
+}
|
|
+
|
|
+static void xgene_enet_write_ring_state(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ int i;
|
|
+
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_CONFIG, ring->num);
|
|
+ for (i = 0; i < NUM_RING_CONFIG; i++) {
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_WR_BASE + (i * 4),
|
|
+ ring->state[i]);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_enet_clr_ring_state(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ memset(ring->state, 0, sizeof(u32) * NUM_RING_CONFIG);
|
|
+ xgene_enet_write_ring_state(ring);
|
|
+}
|
|
+
|
|
+static void xgene_enet_set_ring_state(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ xgene_enet_ring_set_type(ring);
|
|
+
|
|
+ if (xgene_enet_ring_owner(ring->id) == RING_OWNER_ETH0)
|
|
+ xgene_enet_ring_set_recombbuf(ring);
|
|
+
|
|
+ xgene_enet_ring_init(ring);
|
|
+ xgene_enet_write_ring_state(ring);
|
|
+}
|
|
+
|
|
+static void xgene_enet_set_ring_id(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 ring_id_val, ring_id_buf;
|
|
+ bool is_bufpool;
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+
|
|
+ ring_id_val = ring->id & GENMASK(9, 0);
|
|
+ ring_id_val |= OVERWRITE;
|
|
+
|
|
+ ring_id_buf = (ring->num << 9) & GENMASK(18, 9);
|
|
+ ring_id_buf |= PREFETCH_BUF_EN;
|
|
+ if (is_bufpool)
|
|
+ ring_id_buf |= IS_BUFFER_POOL;
|
|
+
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID, ring_id_val);
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID_BUF, ring_id_buf);
|
|
+}
|
|
+
|
|
+static void xgene_enet_clr_desc_ring_id(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 ring_id;
|
|
+
|
|
+ ring_id = ring->id | OVERWRITE;
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID, ring_id);
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_ID_BUF, 0);
|
|
+}
|
|
+
|
|
+struct xgene_enet_desc_ring *xgene_enet_setup_ring(
|
|
+ struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 size = ring->size;
|
|
+ u32 i, data;
|
|
+ u64 *desc;
|
|
+ bool is_bufpool;
|
|
+
|
|
+ xgene_enet_clr_ring_state(ring);
|
|
+ xgene_enet_set_ring_state(ring);
|
|
+ xgene_enet_set_ring_id(ring);
|
|
+
|
|
+ ring->slots = xgene_enet_get_numslots(ring->id, size);
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+ if (is_bufpool || xgene_enet_ring_owner(ring->id) != RING_OWNER_CPU)
|
|
+ return ring;
|
|
+
|
|
+ for (i = 0; i < ring->slots; i++) {
|
|
+ desc = (u64 *)&ring->raw_desc[i];
|
|
+ desc[EMPTY_SLOT_INDEX] = EMPTY_SLOT;
|
|
+ }
|
|
+
|
|
+ xgene_enet_ring_rd32(ring, CSR_RING_NE_INT_MODE, &data);
|
|
+ data |= BIT(31 - xgene_enet_ring_bufnum(ring->id));
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_NE_INT_MODE, data);
|
|
+
|
|
+ return ring;
|
|
+}
|
|
+
|
|
+void xgene_enet_clear_ring(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 data;
|
|
+ bool is_bufpool;
|
|
+
|
|
+ is_bufpool = xgene_enet_is_bufpool(ring->id);
|
|
+ if (is_bufpool || xgene_enet_ring_owner(ring->id) != RING_OWNER_CPU)
|
|
+ goto out;
|
|
+
|
|
+ xgene_enet_ring_rd32(ring, CSR_RING_NE_INT_MODE, &data);
|
|
+ data &= ~BIT(31 - xgene_enet_ring_bufnum(ring->id));
|
|
+ xgene_enet_ring_wr32(ring, CSR_RING_NE_INT_MODE, data);
|
|
+
|
|
+out:
|
|
+ xgene_enet_clr_desc_ring_id(ring);
|
|
+ xgene_enet_clr_ring_state(ring);
|
|
+}
|
|
+
|
|
+void xgene_enet_parse_error(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_pdata *pdata,
|
|
+ enum xgene_enet_err_code status)
|
|
+{
|
|
+ struct rtnl_link_stats64 *stats = &pdata->stats;
|
|
+
|
|
+ switch (status) {
|
|
+ case INGRESS_CRC:
|
|
+ stats->rx_crc_errors++;
|
|
+ break;
|
|
+ case INGRESS_CHECKSUM:
|
|
+ case INGRESS_CHECKSUM_COMPUTE:
|
|
+ stats->rx_errors++;
|
|
+ break;
|
|
+ case INGRESS_TRUNC_FRAME:
|
|
+ stats->rx_frame_errors++;
|
|
+ break;
|
|
+ case INGRESS_PKT_LEN:
|
|
+ stats->rx_length_errors++;
|
|
+ break;
|
|
+ case INGRESS_PKT_UNDER:
|
|
+ stats->rx_frame_errors++;
|
|
+ break;
|
|
+ case INGRESS_FIFO_OVERRUN:
|
|
+ stats->rx_fifo_errors++;
|
|
+ break;
|
|
+ default:
|
|
+ break;
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_csr_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_ring_if(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_ring_if_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_diag_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_diag_csr_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_mcx_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 val)
|
|
+{
|
|
+ void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
|
|
+
|
|
+ iowrite32(val, addr);
|
|
+}
|
|
+
|
|
+static bool xgene_enet_wr_indirect(void __iomem *addr, void __iomem *wr,
|
|
+ void __iomem *cmd, void __iomem *cmd_done,
|
|
+ u32 wr_addr, u32 wr_data)
|
|
+{
|
|
+ u32 done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ iowrite32(wr_addr, addr);
|
|
+ iowrite32(wr_data, wr);
|
|
+ iowrite32(XGENE_ENET_WR_CMD, cmd);
|
|
+
|
|
+ /* wait for write command to complete */
|
|
+ while (!(done = ioread32(cmd_done)) && wait--)
|
|
+ udelay(1);
|
|
+
|
|
+ if (!done)
|
|
+ return false;
|
|
+
|
|
+ iowrite32(0, cmd);
|
|
+
|
|
+ return true;
|
|
+}
|
|
+
|
|
+static void xgene_enet_wr_mcx_mac(struct xgene_enet_pdata *pdata,
|
|
+ u32 wr_addr, u32 wr_data)
|
|
+{
|
|
+ void __iomem *addr, *wr, *cmd, *cmd_done;
|
|
+ bool ret;
|
|
+
|
|
+ addr = pdata->mcx_mac_addr + MAC_ADDR_REG_OFFSET;
|
|
+ wr = pdata->mcx_mac_addr + MAC_WRITE_REG_OFFSET;
|
|
+ cmd = pdata->mcx_mac_addr + MAC_COMMAND_REG_OFFSET;
|
|
+ cmd_done = pdata->mcx_mac_addr + MAC_COMMAND_DONE_REG_OFFSET;
|
|
+
|
|
+ ret = xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data);
|
|
+ if (!ret)
|
|
+ netdev_err(pdata->ndev, "MCX mac write failed, addr: %04x\n",
|
|
+ wr_addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 *val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_csr_addr + offset;
|
|
+
|
|
+ *val = ioread32(addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_diag_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 *val)
|
|
+{
|
|
+ void __iomem *addr = pdata->eth_diag_csr_addr + offset;
|
|
+
|
|
+ *val = ioread32(addr);
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_mcx_csr(struct xgene_enet_pdata *pdata,
|
|
+ u32 offset, u32 *val)
|
|
+{
|
|
+ void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
|
|
+
|
|
+ *val = ioread32(addr);
|
|
+}
|
|
+
|
|
+static bool xgene_enet_rd_indirect(void __iomem *addr, void __iomem *rd,
|
|
+ void __iomem *cmd, void __iomem *cmd_done,
|
|
+ u32 rd_addr, u32 *rd_data)
|
|
+{
|
|
+ u32 done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ iowrite32(rd_addr, addr);
|
|
+ iowrite32(XGENE_ENET_RD_CMD, cmd);
|
|
+
|
|
+ /* wait for read command to complete */
|
|
+ while (!(done = ioread32(cmd_done)) && wait--)
|
|
+ udelay(1);
|
|
+
|
|
+ if (!done)
|
|
+ return false;
|
|
+
|
|
+ *rd_data = ioread32(rd);
|
|
+ iowrite32(0, cmd);
|
|
+
|
|
+ return true;
|
|
+}
|
|
+
|
|
+static void xgene_enet_rd_mcx_mac(struct xgene_enet_pdata *pdata,
|
|
+ u32 rd_addr, u32 *rd_data)
|
|
+{
|
|
+ void __iomem *addr, *rd, *cmd, *cmd_done;
|
|
+ bool ret;
|
|
+
|
|
+ addr = pdata->mcx_mac_addr + MAC_ADDR_REG_OFFSET;
|
|
+ rd = pdata->mcx_mac_addr + MAC_READ_REG_OFFSET;
|
|
+ cmd = pdata->mcx_mac_addr + MAC_COMMAND_REG_OFFSET;
|
|
+ cmd_done = pdata->mcx_mac_addr + MAC_COMMAND_DONE_REG_OFFSET;
|
|
+
|
|
+ ret = xgene_enet_rd_indirect(addr, rd, cmd, cmd_done, rd_addr, rd_data);
|
|
+ if (!ret)
|
|
+ netdev_err(pdata->ndev, "MCX mac read failed, addr: %04x\n",
|
|
+ rd_addr);
|
|
+}
|
|
+
|
|
+static int xgene_mii_phy_write(struct xgene_enet_pdata *pdata, int phy_id,
|
|
+ u32 reg, u16 data)
|
|
+{
|
|
+ u32 addr = 0, wr_data = 0;
|
|
+ u32 done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ PHY_ADDR_SET(&addr, phy_id);
|
|
+ REG_ADDR_SET(&addr, reg);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_ADDRESS_ADDR, addr);
|
|
+
|
|
+ PHY_CONTROL_SET(&wr_data, data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_CONTROL_ADDR, wr_data);
|
|
+ do {
|
|
+ usleep_range(5, 10);
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_INDICATORS_ADDR, &done);
|
|
+ } while ((done & BUSY_MASK) && wait--);
|
|
+
|
|
+ if (done & BUSY_MASK) {
|
|
+ netdev_err(pdata->ndev, "MII_MGMT write failed\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int xgene_mii_phy_read(struct xgene_enet_pdata *pdata,
|
|
+ u8 phy_id, u32 reg)
|
|
+{
|
|
+ u32 addr = 0;
|
|
+ u32 data, done;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ PHY_ADDR_SET(&addr, phy_id);
|
|
+ REG_ADDR_SET(&addr, reg);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_ADDRESS_ADDR, addr);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_COMMAND_ADDR, READ_CYCLE_MASK);
|
|
+ do {
|
|
+ usleep_range(5, 10);
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_INDICATORS_ADDR, &done);
|
|
+ } while ((done & BUSY_MASK) && wait--);
|
|
+
|
|
+ if (done & BUSY_MASK) {
|
|
+ netdev_err(pdata->ndev, "MII_MGMT read failed\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_STATUS_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_COMMAND_ADDR, 0);
|
|
+
|
|
+ return data;
|
|
+}
|
|
+
|
|
+void xgene_gmac_set_mac_addr(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 addr0, addr1;
|
|
+ u8 *dev_addr = pdata->ndev->dev_addr;
|
|
+
|
|
+ addr0 = (dev_addr[3] << 24) | (dev_addr[2] << 16) |
|
|
+ (dev_addr[1] << 8) | dev_addr[0];
|
|
+ addr1 = (dev_addr[5] << 24) | (dev_addr[4] << 16);
|
|
+ addr1 |= pdata->phy_addr & 0xFFFF;
|
|
+
|
|
+ xgene_enet_wr_mcx_mac(pdata, STATION_ADDR0_ADDR, addr0);
|
|
+ xgene_enet_wr_mcx_mac(pdata, STATION_ADDR1_ADDR, addr1);
|
|
+}
|
|
+
|
|
+static int xgene_enet_ecc_init(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct net_device *ndev = pdata->ndev;
|
|
+ u32 data;
|
|
+ u8 wait = 10;
|
|
+
|
|
+ xgene_enet_wr_diag_csr(pdata, ENET_CFG_MEM_RAM_SHUTDOWN_ADDR, 0x0);
|
|
+ do {
|
|
+ usleep_range(100, 110);
|
|
+ xgene_enet_rd_diag_csr(pdata, ENET_BLOCK_MEM_RDY_ADDR, &data);
|
|
+ } while ((data != 0xffffffff) && wait--);
|
|
+
|
|
+ if (data != 0xffffffff) {
|
|
+ netdev_err(ndev, "Failed to release memory from shutdown\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+void xgene_gmac_reset(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, SOFT_RESET1);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, 0);
|
|
+}
|
|
+
|
|
+void xgene_gmac_init(struct xgene_enet_pdata *pdata, int speed)
|
|
+{
|
|
+ u32 value, mc2;
|
|
+ u32 intf_ctl, rgmii;
|
|
+ u32 icm0, icm2;
|
|
+
|
|
+ xgene_gmac_reset(pdata);
|
|
+
|
|
+ xgene_enet_rd_mcx_csr(pdata, ICM_CONFIG0_REG_0_ADDR, &icm0);
|
|
+ xgene_enet_rd_mcx_csr(pdata, ICM_CONFIG2_REG_0_ADDR, &icm2);
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_2_ADDR, &mc2);
|
|
+ xgene_enet_rd_mcx_mac(pdata, INTERFACE_CONTROL_ADDR, &intf_ctl);
|
|
+ xgene_enet_rd_csr(pdata, RGMII_REG_0_ADDR, &rgmii);
|
|
+
|
|
+ switch (speed) {
|
|
+ case SPEED_10:
|
|
+ ENET_INTERFACE_MODE2_SET(&mc2, 1);
|
|
+ CFG_MACMODE_SET(&icm0, 0);
|
|
+ CFG_WAITASYNCRD_SET(&icm2, 500);
|
|
+ rgmii &= ~CFG_SPEED_1250;
|
|
+ break;
|
|
+ case SPEED_100:
|
|
+ ENET_INTERFACE_MODE2_SET(&mc2, 1);
|
|
+ intf_ctl |= ENET_LHD_MODE;
|
|
+ CFG_MACMODE_SET(&icm0, 1);
|
|
+ CFG_WAITASYNCRD_SET(&icm2, 80);
|
|
+ rgmii &= ~CFG_SPEED_1250;
|
|
+ break;
|
|
+ default:
|
|
+ ENET_INTERFACE_MODE2_SET(&mc2, 2);
|
|
+ intf_ctl |= ENET_GHD_MODE;
|
|
+ CFG_TXCLK_MUXSEL0_SET(&rgmii, 4);
|
|
+ xgene_enet_rd_csr(pdata, DEBUG_REG_ADDR, &value);
|
|
+ value |= CFG_BYPASS_UNISEC_TX | CFG_BYPASS_UNISEC_RX;
|
|
+ xgene_enet_wr_csr(pdata, DEBUG_REG_ADDR, value);
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ mc2 |= FULL_DUPLEX2;
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_2_ADDR, mc2);
|
|
+ xgene_enet_wr_mcx_mac(pdata, INTERFACE_CONTROL_ADDR, intf_ctl);
|
|
+
|
|
+ xgene_gmac_set_mac_addr(pdata);
|
|
+
|
|
+ /* Adjust MDC clock frequency */
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, &value);
|
|
+ MGMT_CLOCK_SEL_SET(&value, 7);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, value);
|
|
+
|
|
+ /* Enable drop if bufpool not available */
|
|
+ xgene_enet_rd_csr(pdata, RSIF_CONFIG_REG_ADDR, &value);
|
|
+ value |= CFG_RSIF_FPBUFF_TIMEOUT_EN;
|
|
+ xgene_enet_wr_csr(pdata, RSIF_CONFIG_REG_ADDR, value);
|
|
+
|
|
+ /* Rtype should be copied from FP */
|
|
+ xgene_enet_wr_csr(pdata, RSIF_RAM_DBG_REG0_ADDR, 0);
|
|
+ xgene_enet_wr_csr(pdata, RGMII_REG_0_ADDR, rgmii);
|
|
+
|
|
+ /* Rx-Tx traffic resume */
|
|
+ xgene_enet_wr_csr(pdata, CFG_LINK_AGGR_RESUME_0_ADDR, TX_PORT0);
|
|
+
|
|
+ xgene_enet_wr_mcx_csr(pdata, ICM_CONFIG0_REG_0_ADDR, icm0);
|
|
+ xgene_enet_wr_mcx_csr(pdata, ICM_CONFIG2_REG_0_ADDR, icm2);
|
|
+
|
|
+ xgene_enet_rd_mcx_csr(pdata, RX_DV_GATE_REG_0_ADDR, &value);
|
|
+ value &= ~TX_DV_GATE_EN0;
|
|
+ value &= ~RX_DV_GATE_EN0;
|
|
+ value |= RESUME_RX0;
|
|
+ xgene_enet_wr_mcx_csr(pdata, RX_DV_GATE_REG_0_ADDR, value);
|
|
+
|
|
+ xgene_enet_wr_csr(pdata, CFG_BYPASS_ADDR, RESUME_TX);
|
|
+}
|
|
+
|
|
+static void xgene_enet_config_ring_if_assoc(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 val = 0xffffffff;
|
|
+
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIWQASSOC_ADDR, val);
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIFPQASSOC_ADDR, val);
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEWQASSOC_ADDR, val);
|
|
+ xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEFPQASSOC_ADDR, val);
|
|
+}
|
|
+
|
|
+void xgene_enet_cle_bypass(struct xgene_enet_pdata *pdata,
|
|
+ u32 dst_ring_num, u16 bufpool_id)
|
|
+{
|
|
+ u32 cb;
|
|
+ u32 fpsel;
|
|
+
|
|
+ fpsel = xgene_enet_ring_bufnum(bufpool_id) - 0x20;
|
|
+
|
|
+ xgene_enet_rd_csr(pdata, CLE_BYPASS_REG0_0_ADDR, &cb);
|
|
+ cb |= CFG_CLE_BYPASS_EN0;
|
|
+ CFG_CLE_IP_PROTOCOL0_SET(&cb, 3);
|
|
+ xgene_enet_wr_csr(pdata, CLE_BYPASS_REG0_0_ADDR, cb);
|
|
+
|
|
+ xgene_enet_rd_csr(pdata, CLE_BYPASS_REG1_0_ADDR, &cb);
|
|
+ CFG_CLE_DSTQID0_SET(&cb, dst_ring_num);
|
|
+ CFG_CLE_FPSEL0_SET(&cb, fpsel);
|
|
+ xgene_enet_wr_csr(pdata, CLE_BYPASS_REG1_0_ADDR, cb);
|
|
+}
|
|
+
|
|
+void xgene_gmac_rx_enable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data | RX_EN);
|
|
+}
|
|
+
|
|
+void xgene_gmac_tx_enable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data | TX_EN);
|
|
+}
|
|
+
|
|
+void xgene_gmac_rx_disable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data & ~RX_EN);
|
|
+}
|
|
+
|
|
+void xgene_gmac_tx_disable(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 data;
|
|
+
|
|
+ xgene_enet_rd_mcx_mac(pdata, MAC_CONFIG_1_ADDR, &data);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MAC_CONFIG_1_ADDR, data & ~TX_EN);
|
|
+}
|
|
+
|
|
+void xgene_enet_reset(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ u32 val;
|
|
+
|
|
+ clk_prepare_enable(pdata->clk);
|
|
+ clk_disable_unprepare(pdata->clk);
|
|
+ clk_prepare_enable(pdata->clk);
|
|
+ xgene_enet_ecc_init(pdata);
|
|
+ xgene_enet_config_ring_if_assoc(pdata);
|
|
+
|
|
+ /* Enable auto-incr for scanning */
|
|
+ xgene_enet_rd_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, &val);
|
|
+ val |= SCAN_AUTO_INCR;
|
|
+ MGMT_CLOCK_SEL_SET(&val, 1);
|
|
+ xgene_enet_wr_mcx_mac(pdata, MII_MGMT_CONFIG_ADDR, val);
|
|
+}
|
|
+
|
|
+void xgene_gport_shutdown(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ clk_disable_unprepare(pdata->clk);
|
|
+}
|
|
+
|
|
+static int xgene_enet_mdio_read(struct mii_bus *bus, int mii_id, int regnum)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = bus->priv;
|
|
+ u32 val;
|
|
+
|
|
+ val = xgene_mii_phy_read(pdata, mii_id, regnum);
|
|
+ netdev_dbg(pdata->ndev, "mdio_rd: bus=%d reg=%d val=%x\n",
|
|
+ mii_id, regnum, val);
|
|
+
|
|
+ return val;
|
|
+}
|
|
+
|
|
+static int xgene_enet_mdio_write(struct mii_bus *bus, int mii_id, int regnum,
|
|
+ u16 val)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = bus->priv;
|
|
+ int ret;
|
|
+
|
|
+ netdev_dbg(pdata->ndev, "mdio_wr: bus=%d reg=%d val=%x\n",
|
|
+ mii_id, regnum, val);
|
|
+ ret = xgene_mii_phy_write(pdata, mii_id, regnum, val);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void xgene_enet_adjust_link(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct phy_device *phydev = pdata->phy_dev;
|
|
+
|
|
+ if (phydev->link) {
|
|
+ if (pdata->phy_speed != phydev->speed) {
|
|
+ xgene_gmac_init(pdata, phydev->speed);
|
|
+ xgene_gmac_rx_enable(pdata);
|
|
+ xgene_gmac_tx_enable(pdata);
|
|
+ pdata->phy_speed = phydev->speed;
|
|
+ phy_print_status(phydev);
|
|
+ }
|
|
+ } else {
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+ pdata->phy_speed = SPEED_UNKNOWN;
|
|
+ phy_print_status(phydev);
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_enet_phy_connect(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device_node *phy_np;
|
|
+ struct phy_device *phy_dev;
|
|
+ struct device *dev = &pdata->pdev->dev;
|
|
+
|
|
+ phy_np = of_parse_phandle(dev->of_node, "phy-handle", 0);
|
|
+ if (!phy_np) {
|
|
+ netdev_dbg(ndev, "No phy-handle found\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ phy_dev = of_phy_connect(ndev, phy_np, &xgene_enet_adjust_link,
|
|
+ 0, pdata->phy_mode);
|
|
+ if (!phy_dev) {
|
|
+ netdev_err(ndev, "Could not connect to PHY\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ pdata->phy_speed = SPEED_UNKNOWN;
|
|
+ phy_dev->supported &= ~SUPPORTED_10baseT_Half &
|
|
+ ~SUPPORTED_100baseT_Half &
|
|
+ ~SUPPORTED_1000baseT_Half;
|
|
+ phy_dev->advertising = phy_dev->supported;
|
|
+ pdata->phy_dev = phy_dev;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+int xgene_enet_mdio_config(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct net_device *ndev = pdata->ndev;
|
|
+ struct device *dev = &pdata->pdev->dev;
|
|
+ struct device_node *child_np;
|
|
+ struct device_node *mdio_np = NULL;
|
|
+ struct mii_bus *mdio_bus;
|
|
+ int ret;
|
|
+
|
|
+ for_each_child_of_node(dev->of_node, child_np) {
|
|
+ if (of_device_is_compatible(child_np, "apm,xgene-mdio")) {
|
|
+ mdio_np = child_np;
|
|
+ break;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ if (!mdio_np) {
|
|
+ netdev_dbg(ndev, "No mdio node in the dts\n");
|
|
+ return -1;
|
|
+ }
|
|
+
|
|
+ mdio_bus = mdiobus_alloc();
|
|
+ if (!mdio_bus)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ mdio_bus->name = "APM X-Gene MDIO bus";
|
|
+ mdio_bus->read = xgene_enet_mdio_read;
|
|
+ mdio_bus->write = xgene_enet_mdio_write;
|
|
+ snprintf(mdio_bus->id, MII_BUS_ID_SIZE, "%s-%s", "xgene-mii",
|
|
+ ndev->name);
|
|
+
|
|
+ mdio_bus->priv = pdata;
|
|
+ mdio_bus->parent = &ndev->dev;
|
|
+
|
|
+ ret = of_mdiobus_register(mdio_bus, mdio_np);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "Failed to register MDIO bus\n");
|
|
+ goto err;
|
|
+ }
|
|
+ pdata->mdio_bus = mdio_bus;
|
|
+
|
|
+ ret = xgene_enet_phy_connect(ndev);
|
|
+ if (ret)
|
|
+ goto err;
|
|
+
|
|
+ return ret;
|
|
+
|
|
+err:
|
|
+ mdiobus_free(mdio_bus);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+int xgene_enet_mdio_remove(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct mii_bus *mdio_bus;
|
|
+
|
|
+ mdio_bus = pdata->mdio_bus;
|
|
+ mdiobus_unregister(mdio_bus);
|
|
+ mdiobus_free(mdio_bus);
|
|
+ pdata->mdio_bus = NULL;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_hw.h b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.h
|
|
new file mode 100644
|
|
index 0000000..2041313
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_hw.h
|
|
@@ -0,0 +1,375 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#ifndef __XGENE_ENET_HW_H__
|
|
+#define __XGENE_ENET_HW_H__
|
|
+
|
|
+#include "xgene_enet_main.h"
|
|
+
|
|
+struct xgene_enet_pdata;
|
|
+struct xgene_enet_stats;
|
|
+
|
|
+/* clears and then set bits */
|
|
+static inline void xgene_set_bits(u32 *dst, u32 val, u32 start, u32 len)
|
|
+{
|
|
+ u32 end = start + len - 1;
|
|
+ u32 mask = GENMASK(end, start);
|
|
+
|
|
+ *dst &= ~mask;
|
|
+ *dst |= (val << start) & mask;
|
|
+}
|
|
+
|
|
+static inline u32 xgene_get_bits(u32 val, u32 start, u32 end)
|
|
+{
|
|
+ return (val & GENMASK(end, start)) >> start;
|
|
+}
|
|
+
|
|
+#define CSR_RING_ID 0x0008
|
|
+#define OVERWRITE BIT(31)
|
|
+#define IS_BUFFER_POOL BIT(20)
|
|
+#define PREFETCH_BUF_EN BIT(21)
|
|
+#define CSR_RING_ID_BUF 0x000c
|
|
+#define CSR_RING_NE_INT_MODE 0x017c
|
|
+#define CSR_RING_CONFIG 0x006c
|
|
+#define CSR_RING_WR_BASE 0x0070
|
|
+#define NUM_RING_CONFIG 5
|
|
+#define BUFPOOL_MODE 3
|
|
+#define RM3 3
|
|
+#define INC_DEC_CMD_ADDR 0x002c
|
|
+#define UDP_HDR_SIZE 2
|
|
+#define BUF_LEN_CODE_2K 0x5000
|
|
+
|
|
+#define CREATE_MASK(pos, len) GENMASK((pos)+(len)-1, (pos))
|
|
+#define CREATE_MASK_ULL(pos, len) GENMASK_ULL((pos)+(len)-1, (pos))
|
|
+
|
|
+/* Empty slot soft signature */
|
|
+#define EMPTY_SLOT_INDEX 1
|
|
+#define EMPTY_SLOT ~0ULL
|
|
+
|
|
+#define WORK_DESC_SIZE 32
|
|
+#define BUFPOOL_DESC_SIZE 16
|
|
+
|
|
+#define RING_OWNER_MASK GENMASK(9, 6)
|
|
+#define RING_BUFNUM_MASK GENMASK(5, 0)
|
|
+
|
|
+#define SELTHRSH_POS 3
|
|
+#define SELTHRSH_LEN 3
|
|
+#define RINGADDRL_POS 5
|
|
+#define RINGADDRL_LEN 27
|
|
+#define RINGADDRH_POS 0
|
|
+#define RINGADDRH_LEN 6
|
|
+#define RINGSIZE_POS 23
|
|
+#define RINGSIZE_LEN 3
|
|
+#define RINGTYPE_POS 19
|
|
+#define RINGTYPE_LEN 2
|
|
+#define RINGMODE_POS 20
|
|
+#define RINGMODE_LEN 3
|
|
+#define RECOMTIMEOUTL_POS 28
|
|
+#define RECOMTIMEOUTL_LEN 3
|
|
+#define RECOMTIMEOUTH_POS 0
|
|
+#define RECOMTIMEOUTH_LEN 2
|
|
+#define NUMMSGSINQ_POS 1
|
|
+#define NUMMSGSINQ_LEN 16
|
|
+#define ACCEPTLERR BIT(19)
|
|
+#define QCOHERENT BIT(4)
|
|
+#define RECOMBBUF BIT(27)
|
|
+
|
|
+#define BLOCK_ETH_CSR_OFFSET 0x2000
|
|
+#define BLOCK_ETH_RING_IF_OFFSET 0x9000
|
|
+#define BLOCK_ETH_CLKRST_CSR_OFFSET 0xC000
|
|
+#define BLOCK_ETH_DIAG_CSR_OFFSET 0xD000
|
|
+
|
|
+#define BLOCK_ETH_MAC_OFFSET 0x0000
|
|
+#define BLOCK_ETH_STATS_OFFSET 0x0014
|
|
+#define BLOCK_ETH_MAC_CSR_OFFSET 0x2800
|
|
+
|
|
+#define MAC_ADDR_REG_OFFSET 0x00
|
|
+#define MAC_COMMAND_REG_OFFSET 0x04
|
|
+#define MAC_WRITE_REG_OFFSET 0x08
|
|
+#define MAC_READ_REG_OFFSET 0x0c
|
|
+#define MAC_COMMAND_DONE_REG_OFFSET 0x10
|
|
+
|
|
+#define STAT_ADDR_REG_OFFSET 0x00
|
|
+#define STAT_COMMAND_REG_OFFSET 0x04
|
|
+#define STAT_WRITE_REG_OFFSET 0x08
|
|
+#define STAT_READ_REG_OFFSET 0x0c
|
|
+#define STAT_COMMAND_DONE_REG_OFFSET 0x10
|
|
+
|
|
+#define MII_MGMT_CONFIG_ADDR 0x20
|
|
+#define MII_MGMT_COMMAND_ADDR 0x24
|
|
+#define MII_MGMT_ADDRESS_ADDR 0x28
|
|
+#define MII_MGMT_CONTROL_ADDR 0x2c
|
|
+#define MII_MGMT_STATUS_ADDR 0x30
|
|
+#define MII_MGMT_INDICATORS_ADDR 0x34
|
|
+
|
|
+#define BUSY_MASK BIT(0)
|
|
+#define READ_CYCLE_MASK BIT(0)
|
|
+#define PHY_CONTROL_SET(dst, val) xgene_set_bits(dst, val, 0, 16)
|
|
+
|
|
+#define ENET_SPARE_CFG_REG_ADDR 0x0750
|
|
+#define RSIF_CONFIG_REG_ADDR 0x0010
|
|
+#define RSIF_RAM_DBG_REG0_ADDR 0x0048
|
|
+#define RGMII_REG_0_ADDR 0x07e0
|
|
+#define CFG_LINK_AGGR_RESUME_0_ADDR 0x07c8
|
|
+#define DEBUG_REG_ADDR 0x0700
|
|
+#define CFG_BYPASS_ADDR 0x0294
|
|
+#define CLE_BYPASS_REG0_0_ADDR 0x0490
|
|
+#define CLE_BYPASS_REG1_0_ADDR 0x0494
|
|
+#define CFG_RSIF_FPBUFF_TIMEOUT_EN BIT(31)
|
|
+#define RESUME_TX BIT(0)
|
|
+#define CFG_SPEED_1250 BIT(24)
|
|
+#define TX_PORT0 BIT(0)
|
|
+#define CFG_BYPASS_UNISEC_TX BIT(2)
|
|
+#define CFG_BYPASS_UNISEC_RX BIT(1)
|
|
+#define CFG_CLE_BYPASS_EN0 BIT(31)
|
|
+#define CFG_TXCLK_MUXSEL0_SET(dst, val) xgene_set_bits(dst, val, 29, 3)
|
|
+
|
|
+#define CFG_CLE_IP_PROTOCOL0_SET(dst, val) xgene_set_bits(dst, val, 16, 2)
|
|
+#define CFG_CLE_DSTQID0_SET(dst, val) xgene_set_bits(dst, val, 0, 12)
|
|
+#define CFG_CLE_FPSEL0_SET(dst, val) xgene_set_bits(dst, val, 16, 4)
|
|
+#define CFG_MACMODE_SET(dst, val) xgene_set_bits(dst, val, 18, 2)
|
|
+#define CFG_WAITASYNCRD_SET(dst, val) xgene_set_bits(dst, val, 0, 16)
|
|
+#define ICM_CONFIG0_REG_0_ADDR 0x0400
|
|
+#define ICM_CONFIG2_REG_0_ADDR 0x0410
|
|
+#define RX_DV_GATE_REG_0_ADDR 0x05fc
|
|
+#define TX_DV_GATE_EN0 BIT(2)
|
|
+#define RX_DV_GATE_EN0 BIT(1)
|
|
+#define RESUME_RX0 BIT(0)
|
|
+#define ENET_CFGSSQMIWQASSOC_ADDR 0xe0
|
|
+#define ENET_CFGSSQMIFPQASSOC_ADDR 0xdc
|
|
+#define ENET_CFGSSQMIQMLITEFPQASSOC_ADDR 0xf0
|
|
+#define ENET_CFGSSQMIQMLITEWQASSOC_ADDR 0xf4
|
|
+#define ENET_CFG_MEM_RAM_SHUTDOWN_ADDR 0x70
|
|
+#define ENET_BLOCK_MEM_RDY_ADDR 0x74
|
|
+#define MAC_CONFIG_1_ADDR 0x00
|
|
+#define MAC_CONFIG_2_ADDR 0x04
|
|
+#define MAX_FRAME_LEN_ADDR 0x10
|
|
+#define INTERFACE_CONTROL_ADDR 0x38
|
|
+#define STATION_ADDR0_ADDR 0x40
|
|
+#define STATION_ADDR1_ADDR 0x44
|
|
+#define PHY_ADDR_SET(dst, val) xgene_set_bits(dst, val, 8, 5)
|
|
+#define REG_ADDR_SET(dst, val) xgene_set_bits(dst, val, 0, 5)
|
|
+#define ENET_INTERFACE_MODE2_SET(dst, val) xgene_set_bits(dst, val, 8, 2)
|
|
+#define MGMT_CLOCK_SEL_SET(dst, val) xgene_set_bits(dst, val, 0, 3)
|
|
+#define SOFT_RESET1 BIT(31)
|
|
+#define TX_EN BIT(0)
|
|
+#define RX_EN BIT(2)
|
|
+#define ENET_LHD_MODE BIT(25)
|
|
+#define ENET_GHD_MODE BIT(26)
|
|
+#define FULL_DUPLEX2 BIT(0)
|
|
+#define SCAN_AUTO_INCR BIT(5)
|
|
+#define TBYT_ADDR 0x38
|
|
+#define TPKT_ADDR 0x39
|
|
+#define TDRP_ADDR 0x45
|
|
+#define TFCS_ADDR 0x47
|
|
+#define TUND_ADDR 0x4a
|
|
+
|
|
+#define TSO_IPPROTO_TCP 1
|
|
+#define FULL_DUPLEX 2
|
|
+
|
|
+#define USERINFO_POS 0
|
|
+#define USERINFO_LEN 32
|
|
+#define FPQNUM_POS 32
|
|
+#define FPQNUM_LEN 12
|
|
+#define LERR_POS 60
|
|
+#define LERR_LEN 3
|
|
+#define STASH_POS 52
|
|
+#define STASH_LEN 2
|
|
+#define BUFDATALEN_POS 48
|
|
+#define BUFDATALEN_LEN 12
|
|
+#define DATAADDR_POS 0
|
|
+#define DATAADDR_LEN 42
|
|
+#define COHERENT_POS 63
|
|
+#define HENQNUM_POS 48
|
|
+#define HENQNUM_LEN 12
|
|
+#define TYPESEL_POS 44
|
|
+#define TYPESEL_LEN 4
|
|
+#define ETHHDR_POS 12
|
|
+#define IC_POS 35 /* Insert CRC */
|
|
+#define TCPHDR_POS 0
|
|
+#define TCPHDR_LEN 6
|
|
+#define IPHDR_POS 6
|
|
+#define IPHDR_LEN 6
|
|
+#define EC_POS 22 /* Enable checksum */
|
|
+#define IS_POS 24 /* IP protocol select */
|
|
+
|
|
+#define DATAADDR_MASK CREATE_MASK_ULL(DATAADDR_POS, DATAADDR_LEN)
|
|
+#define BUFDATALEN_MASK CREATE_MASK_ULL(BUFDATALEN_POS, BUFDATALEN_LEN)
|
|
+#define USERINFO_MASK CREATE_MASK_ULL(USERINFO_POS, USERINFO_LEN)
|
|
+#define FPQNUM_MASK CREATE_MASK_ULL(FPQNUM_POS, FPQNUM_LEN)
|
|
+#define LERR_MASK CREATE_MASK_ULL(LERR_POS, LERR_LEN)
|
|
+#define STASHING_MASK CREATE_MASK_ULL(STASH_POS, STASH_LEN)
|
|
+#define COHERENT_MASK BIT_ULL(COHERENT_POS)
|
|
+#define HENQNUM_MASK CREATE_MASK_ULL(HENQNUM_POS, HENQNUM_LEN)
|
|
+#define TCPHDR_MASK CREATE_MASK(TCPHDR_POS, TCPHDR_LEN)
|
|
+#define IPHDR_MASK CREATE_MASK(IPHDR_POS, IPHDR_LEN)
|
|
+#define EC_MASK BIT(EC_POS)
|
|
+#define IS_MASK BIT(IS_POS)
|
|
+#define INSERT_CRC BIT_ULL(IC_POS)
|
|
+#define TYPE_ETH_WORK_MESSAGE BIT_ULL(44)
|
|
+
|
|
+struct xgene_enet_raw_desc {
|
|
+ u64 m0;
|
|
+ u64 m1;
|
|
+ u64 m2;
|
|
+ u64 m3;
|
|
+};
|
|
+
|
|
+struct xgene_enet_raw_desc16 {
|
|
+ u64 m0;
|
|
+ u64 m1;
|
|
+};
|
|
+
|
|
+static inline void xgene_enet_cpu_to_le64(void *desc_ptr, int count)
|
|
+{
|
|
+ u64 *desc = desc_ptr;
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < count; i++)
|
|
+ desc[i] = cpu_to_le64(desc[i]);
|
|
+}
|
|
+
|
|
+static inline void xgene_enet_le64_to_cpu(void *desc_ptr, int count)
|
|
+{
|
|
+ u64 *desc = desc_ptr;
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < count; i++)
|
|
+ desc[i] = le64_to_cpu(desc[i]);
|
|
+}
|
|
+
|
|
+static inline void xgene_enet_desc16_to_le64(void *desc_ptr)
|
|
+{
|
|
+ u64 *desc;
|
|
+
|
|
+ desc = desc_ptr;
|
|
+ desc[1] = cpu_to_le64(desc[1]);
|
|
+}
|
|
+
|
|
+static inline void xgene_enet_le64_to_desc16(void *desc_ptr)
|
|
+{
|
|
+ u64 *desc;
|
|
+
|
|
+ desc = desc_ptr;
|
|
+ desc[1] = le64_to_cpu(desc[1]);
|
|
+}
|
|
+
|
|
+enum xgene_enet_ring_cfgsize {
|
|
+ RING_CFGSIZE_512B,
|
|
+ RING_CFGSIZE_2KB,
|
|
+ RING_CFGSIZE_16KB,
|
|
+ RING_CFGSIZE_64KB,
|
|
+ RING_CFGSIZE_512KB,
|
|
+ RING_CFGSIZE_INVALID
|
|
+};
|
|
+
|
|
+enum xgene_enet_ring_type {
|
|
+ RING_DISABLED,
|
|
+ RING_REGULAR,
|
|
+ RING_BUFPOOL
|
|
+};
|
|
+
|
|
+enum xgene_ring_owner {
|
|
+ RING_OWNER_ETH0,
|
|
+ RING_OWNER_CPU = 15,
|
|
+ RING_OWNER_INVALID
|
|
+};
|
|
+
|
|
+enum xgene_enet_ring_bufnum {
|
|
+ RING_BUFNUM_REGULAR = 0x0,
|
|
+ RING_BUFNUM_BUFPOOL = 0x20,
|
|
+ RING_BUFNUM_INVALID
|
|
+};
|
|
+
|
|
+enum xgene_enet_cmd {
|
|
+ XGENE_ENET_WR_CMD = BIT(31),
|
|
+ XGENE_ENET_RD_CMD = BIT(30)
|
|
+};
|
|
+
|
|
+enum xgene_enet_err_code {
|
|
+ HBF_READ_DATA = 3,
|
|
+ HBF_LL_READ = 4,
|
|
+ BAD_WORK_MSG = 6,
|
|
+ BUFPOOL_TIMEOUT = 15,
|
|
+ INGRESS_CRC = 16,
|
|
+ INGRESS_CHECKSUM = 17,
|
|
+ INGRESS_TRUNC_FRAME = 18,
|
|
+ INGRESS_PKT_LEN = 19,
|
|
+ INGRESS_PKT_UNDER = 20,
|
|
+ INGRESS_FIFO_OVERRUN = 21,
|
|
+ INGRESS_CHECKSUM_COMPUTE = 26,
|
|
+ ERR_CODE_INVALID
|
|
+};
|
|
+
|
|
+static inline enum xgene_ring_owner xgene_enet_ring_owner(u16 id)
|
|
+{
|
|
+ return (id & RING_OWNER_MASK) >> 6;
|
|
+}
|
|
+
|
|
+static inline u8 xgene_enet_ring_bufnum(u16 id)
|
|
+{
|
|
+ return id & RING_BUFNUM_MASK;
|
|
+}
|
|
+
|
|
+static inline bool xgene_enet_is_bufpool(u16 id)
|
|
+{
|
|
+ return ((id & RING_BUFNUM_MASK) >= 0x20) ? true : false;
|
|
+}
|
|
+
|
|
+static inline u16 xgene_enet_get_numslots(u16 id, u32 size)
|
|
+{
|
|
+ bool is_bufpool = xgene_enet_is_bufpool(id);
|
|
+
|
|
+ return (is_bufpool) ? size / BUFPOOL_DESC_SIZE :
|
|
+ size / WORK_DESC_SIZE;
|
|
+}
|
|
+
|
|
+struct xgene_enet_desc_ring *xgene_enet_setup_ring(
|
|
+ struct xgene_enet_desc_ring *ring);
|
|
+void xgene_enet_clear_ring(struct xgene_enet_desc_ring *ring);
|
|
+
|
|
+void xgene_set_tx_desc(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc);
|
|
+void xgene_get_desc(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc);
|
|
+void xgene_get_bufpool_desc(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_raw_desc16 *raw_desc);
|
|
+void xgene_enet_parse_error(struct xgene_enet_desc_ring *ring,
|
|
+ struct xgene_enet_pdata *pdata,
|
|
+ enum xgene_enet_err_code status);
|
|
+
|
|
+void xgene_enet_reset(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_reset(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_init(struct xgene_enet_pdata *priv, int speed);
|
|
+void xgene_gmac_tx_enable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_rx_enable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_tx_disable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_rx_disable(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_set_mac_addr(struct xgene_enet_pdata *pdata);
|
|
+void xgene_enet_cle_bypass(struct xgene_enet_pdata *pdata,
|
|
+ u32 dst_ring_num, u16 bufpool_id);
|
|
+void xgene_gport_shutdown(struct xgene_enet_pdata *priv);
|
|
+void xgene_gmac_get_tx_stats(struct xgene_enet_pdata *pdata);
|
|
+
|
|
+int xgene_enet_mdio_config(struct xgene_enet_pdata *pdata);
|
|
+int xgene_enet_mdio_remove(struct xgene_enet_pdata *pdata);
|
|
+
|
|
+#endif /* __XGENE_ENET_HW_H__ */
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_main.c b/drivers/net/ethernet/apm/xgene/xgene_enet_main.c
|
|
new file mode 100644
|
|
index 0000000..756523a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_main.c
|
|
@@ -0,0 +1,962 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#include "xgene_enet_main.h"
|
|
+#include "xgene_enet_hw.h"
|
|
+
|
|
+static void xgene_enet_init_bufpool(struct xgene_enet_desc_ring *buf_pool)
|
|
+{
|
|
+ struct xgene_enet_raw_desc16 *raw_desc;
|
|
+ int i;
|
|
+
|
|
+ for (i = 0; i < buf_pool->slots; i++) {
|
|
+ raw_desc = &buf_pool->raw_desc16[i];
|
|
+
|
|
+ /* Hardware expects descriptor in little endian format */
|
|
+ raw_desc->m0 = cpu_to_le64(i |
|
|
+ (((u64)buf_pool->dst_ring_num << FPQNUM_POS) &
|
|
+ FPQNUM_MASK) | STASHING_MASK);
|
|
+ }
|
|
+}
|
|
+
|
|
+static struct device *ndev_to_dev(struct net_device *ndev)
|
|
+{
|
|
+ return ndev->dev.parent;
|
|
+}
|
|
+
|
|
+static int xgene_enet_refill_bufpool(struct xgene_enet_desc_ring *buf_pool,
|
|
+ u32 nbuf)
|
|
+{
|
|
+ struct sk_buff *skb;
|
|
+ struct xgene_enet_raw_desc16 *raw_desc;
|
|
+ struct net_device *ndev;
|
|
+ struct device *dev;
|
|
+ dma_addr_t dma_addr;
|
|
+ u32 tail = buf_pool->tail;
|
|
+ u32 slots = buf_pool->slots - 1;
|
|
+ u16 bufdatalen, len;
|
|
+ int i;
|
|
+
|
|
+ ndev = buf_pool->ndev;
|
|
+ dev = ndev_to_dev(buf_pool->ndev);
|
|
+ bufdatalen = BUF_LEN_CODE_2K | (SKB_BUFFER_SIZE & GENMASK(11, 0));
|
|
+ len = XGENE_ENET_MAX_MTU;
|
|
+
|
|
+ for (i = 0; i < nbuf; i++) {
|
|
+ raw_desc = &buf_pool->raw_desc16[tail];
|
|
+
|
|
+ skb = netdev_alloc_skb_ip_align(ndev, len);
|
|
+ if (unlikely(!skb))
|
|
+ return -ENOMEM;
|
|
+ buf_pool->rx_skb[tail] = skb;
|
|
+
|
|
+ dma_addr = dma_map_single(dev, skb->data, len, DMA_FROM_DEVICE);
|
|
+ if (dma_mapping_error(dev, dma_addr)) {
|
|
+ netdev_err(ndev, "DMA mapping error\n");
|
|
+ dev_kfree_skb_any(skb);
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ raw_desc->m1 = cpu_to_le64((dma_addr & DATAADDR_MASK) |
|
|
+ (((u64)bufdatalen << BUFDATALEN_POS) &
|
|
+ BUFDATALEN_MASK) | COHERENT_MASK);
|
|
+ tail = (tail + 1) & slots;
|
|
+ }
|
|
+
|
|
+ iowrite32(nbuf, buf_pool->cmd);
|
|
+ buf_pool->tail = tail;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static u16 xgene_enet_dst_ring_num(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+
|
|
+ return ((u16)pdata->rm << 10) | ring->num;
|
|
+}
|
|
+
|
|
+static u8 xgene_enet_hdr_len(const void *data)
|
|
+{
|
|
+ const struct ethhdr *eth = data;
|
|
+
|
|
+ return (eth->h_proto == htons(ETH_P_8021Q)) ? VLAN_ETH_HLEN : ETH_HLEN;
|
|
+}
|
|
+
|
|
+static u32 xgene_enet_ring_len(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ u32 *cmd_base = ring->cmd_base;
|
|
+ u32 ring_state, num_msgs;
|
|
+
|
|
+ ring_state = ioread32(&cmd_base[1]);
|
|
+ num_msgs = ring_state & CREATE_MASK(NUMMSGSINQ_POS, NUMMSGSINQ_LEN);
|
|
+
|
|
+ return num_msgs >> NUMMSGSINQ_POS;
|
|
+}
|
|
+
|
|
+static void xgene_enet_delete_bufpool(struct xgene_enet_desc_ring *buf_pool)
|
|
+{
|
|
+ struct xgene_enet_raw_desc16 *raw_desc;
|
|
+ u32 slots = buf_pool->slots - 1;
|
|
+ u32 tail = buf_pool->tail;
|
|
+ u32 userinfo;
|
|
+ int i, len;
|
|
+
|
|
+ len = xgene_enet_ring_len(buf_pool);
|
|
+ for (i = 0; i < len; i++) {
|
|
+ tail = (tail - 1) & slots;
|
|
+ raw_desc = &buf_pool->raw_desc16[tail];
|
|
+
|
|
+ /* Hardware stores descriptor in little endian format */
|
|
+ userinfo = le64_to_cpu(raw_desc->m0) & USERINFO_MASK;
|
|
+ dev_kfree_skb_any(buf_pool->rx_skb[userinfo]);
|
|
+ }
|
|
+
|
|
+ iowrite32(-len, buf_pool->cmd);
|
|
+ buf_pool->tail = tail;
|
|
+}
|
|
+
|
|
+static irqreturn_t xgene_enet_rx_irq(const int irq, void *data)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *rx_ring = data;
|
|
+
|
|
+ if (napi_schedule_prep(&rx_ring->napi)) {
|
|
+ disable_irq_nosync(irq);
|
|
+ __napi_schedule(&rx_ring->napi);
|
|
+ }
|
|
+
|
|
+ return IRQ_HANDLED;
|
|
+}
|
|
+
|
|
+static int xgene_enet_tx_completion(struct xgene_enet_desc_ring *cp_ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc)
|
|
+{
|
|
+ struct sk_buff *skb;
|
|
+ struct device *dev;
|
|
+ u16 skb_index;
|
|
+ u8 status;
|
|
+ int ret = 0;
|
|
+
|
|
+ skb_index = raw_desc->m0 & USERINFO_MASK;
|
|
+ skb = cp_ring->cp_skb[skb_index];
|
|
+
|
|
+ dev = ndev_to_dev(cp_ring->ndev);
|
|
+ dma_unmap_single(dev, raw_desc->m1 & DATAADDR_MASK,
|
|
+ (raw_desc->m1 & BUFDATALEN_MASK) >> BUFDATALEN_POS,
|
|
+ DMA_TO_DEVICE);
|
|
+
|
|
+ /* Checking for error */
|
|
+ status = (raw_desc->m0 & LERR_MASK) >> LERR_POS;
|
|
+ if (unlikely(status > 2)) {
|
|
+ xgene_enet_parse_error(cp_ring, netdev_priv(cp_ring->ndev),
|
|
+ status);
|
|
+ ret = -1;
|
|
+ }
|
|
+
|
|
+ if (likely(skb)) {
|
|
+ dev_kfree_skb_any(skb);
|
|
+ } else {
|
|
+ netdev_err(cp_ring->ndev, "completion skb is NULL\n");
|
|
+ ret = -1;
|
|
+ }
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static u64 xgene_enet_work_msg(struct sk_buff *skb)
|
|
+{
|
|
+ struct iphdr *iph;
|
|
+ u8 l3hlen, l4hlen = 0;
|
|
+ u8 csum_enable = 0;
|
|
+ u8 proto = 0;
|
|
+ u8 ethhdr;
|
|
+ u64 hopinfo;
|
|
+
|
|
+ if (unlikely(skb->protocol != htons(ETH_P_IP)) &&
|
|
+ unlikely(skb->protocol != htons(ETH_P_8021Q)))
|
|
+ goto out;
|
|
+
|
|
+ if (unlikely(!(skb->dev->features & NETIF_F_IP_CSUM)))
|
|
+ goto out;
|
|
+
|
|
+ iph = ip_hdr(skb);
|
|
+ if (unlikely(ip_is_fragment(iph)))
|
|
+ goto out;
|
|
+
|
|
+ if (likely(iph->protocol == IPPROTO_TCP)) {
|
|
+ l4hlen = tcp_hdrlen(skb) >> 2;
|
|
+ csum_enable = 1;
|
|
+ proto = TSO_IPPROTO_TCP;
|
|
+ } else if (iph->protocol == IPPROTO_UDP) {
|
|
+ l4hlen = UDP_HDR_SIZE;
|
|
+ csum_enable = 1;
|
|
+ }
|
|
+out:
|
|
+ l3hlen = ip_hdrlen(skb) >> 2;
|
|
+ ethhdr = xgene_enet_hdr_len(skb->data);
|
|
+ hopinfo = (l4hlen & TCPHDR_MASK) |
|
|
+ ((l3hlen << IPHDR_POS) & IPHDR_MASK) |
|
|
+ (ethhdr << ETHHDR_POS) |
|
|
+ (csum_enable << EC_POS) |
|
|
+ (proto << IS_POS) |
|
|
+ INSERT_CRC |
|
|
+ TYPE_ETH_WORK_MESSAGE;
|
|
+
|
|
+ return hopinfo;
|
|
+}
|
|
+
|
|
+static int xgene_enet_setup_tx_desc(struct xgene_enet_desc_ring *tx_ring,
|
|
+ struct sk_buff *skb)
|
|
+{
|
|
+ struct device *dev = ndev_to_dev(tx_ring->ndev);
|
|
+ struct xgene_enet_raw_desc *raw_desc;
|
|
+ dma_addr_t dma_addr;
|
|
+ u16 tail = tx_ring->tail;
|
|
+ u64 hopinfo;
|
|
+
|
|
+ raw_desc = &tx_ring->raw_desc[tail];
|
|
+ memset(raw_desc, 0, sizeof(struct xgene_enet_raw_desc));
|
|
+
|
|
+ dma_addr = dma_map_single(dev, skb->data, skb->len, DMA_TO_DEVICE);
|
|
+ if (dma_mapping_error(dev, dma_addr)) {
|
|
+ netdev_err(tx_ring->ndev, "DMA mapping error\n");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ /* Hardware expects descriptor in little endian format */
|
|
+ raw_desc->m0 = cpu_to_le64(tail);
|
|
+ raw_desc->m1 = cpu_to_le64((dma_addr & DATAADDR_MASK) |
|
|
+ (((u64)skb->len << BUFDATALEN_POS) & BUFDATALEN_MASK) |
|
|
+ COHERENT_MASK);
|
|
+ hopinfo = xgene_enet_work_msg(skb);
|
|
+ raw_desc->m3 = cpu_to_le64(
|
|
+ (((u64)tx_ring->dst_ring_num << HENQNUM_POS) &
|
|
+ HENQNUM_MASK) | hopinfo);
|
|
+ tx_ring->cp_ring->cp_skb[tail] = skb;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static netdev_tx_t xgene_enet_start_xmit(struct sk_buff *skb,
|
|
+ struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct xgene_enet_desc_ring *tx_ring = pdata->tx_ring;
|
|
+ struct xgene_enet_desc_ring *cp_ring = tx_ring->cp_ring;
|
|
+ u32 tx_level, cq_level;
|
|
+
|
|
+ tx_level = xgene_enet_ring_len(tx_ring);
|
|
+ cq_level = xgene_enet_ring_len(cp_ring);
|
|
+ if (unlikely(tx_level > pdata->tx_qcnt_hi ||
|
|
+ cq_level > pdata->cp_qcnt_hi)) {
|
|
+ netif_stop_queue(ndev);
|
|
+ return NETDEV_TX_BUSY;
|
|
+ }
|
|
+
|
|
+ if (xgene_enet_setup_tx_desc(tx_ring, skb)) {
|
|
+ dev_kfree_skb_any(skb);
|
|
+ return NETDEV_TX_OK;
|
|
+ }
|
|
+
|
|
+ iowrite32(1, tx_ring->cmd);
|
|
+ skb_tx_timestamp(skb);
|
|
+ tx_ring->tail = (tx_ring->tail + 1) & (tx_ring->slots - 1);
|
|
+
|
|
+ pdata->stats.tx_packets++;
|
|
+ pdata->stats.tx_bytes += skb->len;
|
|
+
|
|
+ return NETDEV_TX_OK;
|
|
+}
|
|
+
|
|
+static void xgene_enet_skip_csum(struct sk_buff *skb)
|
|
+{
|
|
+ struct iphdr *iph = ip_hdr(skb);
|
|
+
|
|
+ if (!ip_is_fragment(iph) ||
|
|
+ (iph->protocol != IPPROTO_TCP && iph->protocol != IPPROTO_UDP)) {
|
|
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_enet_rx_frame(struct xgene_enet_desc_ring *rx_ring,
|
|
+ struct xgene_enet_raw_desc *raw_desc)
|
|
+{
|
|
+ struct net_device *ndev;
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev;
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+ u32 datalen, skb_index;
|
|
+ struct sk_buff *skb;
|
|
+ u8 status;
|
|
+ int ret = 0;
|
|
+
|
|
+ ndev = rx_ring->ndev;
|
|
+ pdata = netdev_priv(ndev);
|
|
+ dev = ndev_to_dev(rx_ring->ndev);
|
|
+ buf_pool = rx_ring->buf_pool;
|
|
+
|
|
+ dma_unmap_single(dev, raw_desc->m1 & DATAADDR_MASK, XGENE_ENET_MAX_MTU,
|
|
+ DMA_FROM_DEVICE);
|
|
+ skb_index = raw_desc->m0 & USERINFO_MASK;
|
|
+ skb = buf_pool->rx_skb[skb_index];
|
|
+
|
|
+ /* checking for error */
|
|
+ status = (raw_desc->m0 & LERR_MASK) >> LERR_POS;
|
|
+ if (unlikely(status > 2)) {
|
|
+ dev_kfree_skb_any(skb);
|
|
+ xgene_enet_parse_error(rx_ring, netdev_priv(rx_ring->ndev),
|
|
+ status);
|
|
+ pdata->stats.rx_dropped++;
|
|
+ ret = -1;
|
|
+ goto out;
|
|
+ }
|
|
+
|
|
+ /* strip off CRC as HW isn't doing this */
|
|
+ datalen = (raw_desc->m1 & BUFDATALEN_MASK) >> BUFDATALEN_POS;
|
|
+ datalen -= 4;
|
|
+ prefetch(skb->data - NET_IP_ALIGN);
|
|
+ skb_put(skb, datalen);
|
|
+
|
|
+ skb_checksum_none_assert(skb);
|
|
+ skb->protocol = eth_type_trans(skb, ndev);
|
|
+ if (likely((ndev->features & NETIF_F_IP_CSUM) &&
|
|
+ skb->protocol == htons(ETH_P_IP))) {
|
|
+ xgene_enet_skip_csum(skb);
|
|
+ }
|
|
+
|
|
+ pdata->stats.rx_packets++;
|
|
+ pdata->stats.rx_bytes += datalen;
|
|
+ napi_gro_receive(&rx_ring->napi, skb);
|
|
+out:
|
|
+ if (--rx_ring->nbufpool == 0) {
|
|
+ ret = xgene_enet_refill_bufpool(buf_pool, NUM_BUFPOOL);
|
|
+ rx_ring->nbufpool = NUM_BUFPOOL;
|
|
+ }
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static bool is_rx_desc(struct xgene_enet_raw_desc *raw_desc)
|
|
+{
|
|
+ /* Hardware stores descriptor in little endian format */
|
|
+ raw_desc->m0 = le64_to_cpu(raw_desc->m0);
|
|
+ raw_desc->m1 = le64_to_cpu(raw_desc->m1);
|
|
+ return ((raw_desc->m0 & FPQNUM_MASK) >> FPQNUM_POS) ? true : false;
|
|
+}
|
|
+
|
|
+static int xgene_enet_process_ring(struct xgene_enet_desc_ring *ring,
|
|
+ int budget)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev);
|
|
+ struct xgene_enet_raw_desc *raw_desc;
|
|
+ u16 head = ring->head;
|
|
+ u16 slots = ring->slots - 1;
|
|
+ int ret, count = 0;
|
|
+
|
|
+ do {
|
|
+ raw_desc = &ring->raw_desc[head];
|
|
+ if (unlikely(((u64 *)raw_desc)[EMPTY_SLOT_INDEX] == EMPTY_SLOT))
|
|
+ break;
|
|
+
|
|
+ if (is_rx_desc(raw_desc))
|
|
+ ret = xgene_enet_rx_frame(ring, raw_desc);
|
|
+ else
|
|
+ ret = xgene_enet_tx_completion(ring, raw_desc);
|
|
+ ((u64 *)raw_desc)[EMPTY_SLOT_INDEX] = EMPTY_SLOT;
|
|
+
|
|
+ head = (head + 1) & slots;
|
|
+ count++;
|
|
+
|
|
+ if (ret)
|
|
+ break;
|
|
+ } while (--budget);
|
|
+
|
|
+ if (likely(count)) {
|
|
+ iowrite32(-count, ring->cmd);
|
|
+ ring->head = head;
|
|
+
|
|
+ if (netif_queue_stopped(ring->ndev)) {
|
|
+ if (xgene_enet_ring_len(ring) < pdata->cp_qcnt_low)
|
|
+ netif_wake_queue(ring->ndev);
|
|
+ }
|
|
+ }
|
|
+
|
|
+ return budget;
|
|
+}
|
|
+
|
|
+static int xgene_enet_napi(struct napi_struct *napi, const int budget)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *ring;
|
|
+ int processed;
|
|
+
|
|
+ ring = container_of(napi, struct xgene_enet_desc_ring, napi);
|
|
+ processed = xgene_enet_process_ring(ring, budget);
|
|
+
|
|
+ if (processed != budget) {
|
|
+ napi_complete(napi);
|
|
+ enable_irq(ring->irq);
|
|
+ }
|
|
+
|
|
+ return processed;
|
|
+}
|
|
+
|
|
+static void xgene_enet_timeout(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+
|
|
+ xgene_gmac_reset(pdata);
|
|
+}
|
|
+
|
|
+static int xgene_enet_register_irq(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device *dev = ndev_to_dev(ndev);
|
|
+ int ret;
|
|
+
|
|
+ ret = devm_request_irq(dev, pdata->rx_ring->irq, xgene_enet_rx_irq,
|
|
+ IRQF_SHARED, ndev->name, pdata->rx_ring);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "rx%d interrupt request failed\n",
|
|
+ pdata->rx_ring->irq);
|
|
+ }
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static void xgene_enet_free_irq(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev;
|
|
+
|
|
+ pdata = netdev_priv(ndev);
|
|
+ dev = ndev_to_dev(ndev);
|
|
+ devm_free_irq(dev, pdata->rx_ring->irq, pdata->rx_ring);
|
|
+}
|
|
+
|
|
+static int xgene_enet_open(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ int ret;
|
|
+
|
|
+ xgene_gmac_tx_enable(pdata);
|
|
+ xgene_gmac_rx_enable(pdata);
|
|
+
|
|
+ ret = xgene_enet_register_irq(ndev);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+ napi_enable(&pdata->rx_ring->napi);
|
|
+
|
|
+ if (pdata->phy_dev)
|
|
+ phy_start(pdata->phy_dev);
|
|
+
|
|
+ netif_start_queue(ndev);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_close(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+
|
|
+ netif_stop_queue(ndev);
|
|
+
|
|
+ if (pdata->phy_dev)
|
|
+ phy_stop(pdata->phy_dev);
|
|
+
|
|
+ napi_disable(&pdata->rx_ring->napi);
|
|
+ xgene_enet_free_irq(ndev);
|
|
+ xgene_enet_process_ring(pdata->rx_ring, -1);
|
|
+
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_enet_delete_ring(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev;
|
|
+
|
|
+ pdata = netdev_priv(ring->ndev);
|
|
+ dev = ndev_to_dev(ring->ndev);
|
|
+
|
|
+ xgene_enet_clear_ring(ring);
|
|
+ dma_free_coherent(dev, ring->size, ring->desc_addr, ring->dma);
|
|
+}
|
|
+
|
|
+static void xgene_enet_delete_desc_rings(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+
|
|
+ if (pdata->tx_ring) {
|
|
+ xgene_enet_delete_ring(pdata->tx_ring);
|
|
+ pdata->tx_ring = NULL;
|
|
+ }
|
|
+
|
|
+ if (pdata->rx_ring) {
|
|
+ buf_pool = pdata->rx_ring->buf_pool;
|
|
+ xgene_enet_delete_bufpool(buf_pool);
|
|
+ xgene_enet_delete_ring(buf_pool);
|
|
+ xgene_enet_delete_ring(pdata->rx_ring);
|
|
+ pdata->rx_ring = NULL;
|
|
+ }
|
|
+}
|
|
+
|
|
+static int xgene_enet_get_ring_size(struct device *dev,
|
|
+ enum xgene_enet_ring_cfgsize cfgsize)
|
|
+{
|
|
+ int size = -EINVAL;
|
|
+
|
|
+ switch (cfgsize) {
|
|
+ case RING_CFGSIZE_512B:
|
|
+ size = 0x200;
|
|
+ break;
|
|
+ case RING_CFGSIZE_2KB:
|
|
+ size = 0x800;
|
|
+ break;
|
|
+ case RING_CFGSIZE_16KB:
|
|
+ size = 0x4000;
|
|
+ break;
|
|
+ case RING_CFGSIZE_64KB:
|
|
+ size = 0x10000;
|
|
+ break;
|
|
+ case RING_CFGSIZE_512KB:
|
|
+ size = 0x80000;
|
|
+ break;
|
|
+ default:
|
|
+ dev_err(dev, "Unsupported cfg ring size %d\n", cfgsize);
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ return size;
|
|
+}
|
|
+
|
|
+static void xgene_enet_free_desc_ring(struct xgene_enet_desc_ring *ring)
|
|
+{
|
|
+ struct device *dev;
|
|
+
|
|
+ if (!ring)
|
|
+ return;
|
|
+
|
|
+ dev = ndev_to_dev(ring->ndev);
|
|
+
|
|
+ if (ring->desc_addr) {
|
|
+ xgene_enet_clear_ring(ring);
|
|
+ dma_free_coherent(dev, ring->size, ring->desc_addr, ring->dma);
|
|
+ }
|
|
+ devm_kfree(dev, ring);
|
|
+}
|
|
+
|
|
+static void xgene_enet_free_desc_rings(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct device *dev = &pdata->pdev->dev;
|
|
+ struct xgene_enet_desc_ring *ring;
|
|
+
|
|
+ ring = pdata->tx_ring;
|
|
+ if (ring && ring->cp_ring && ring->cp_ring->cp_skb)
|
|
+ devm_kfree(dev, ring->cp_ring->cp_skb);
|
|
+ xgene_enet_free_desc_ring(ring);
|
|
+
|
|
+ ring = pdata->rx_ring;
|
|
+ if (ring && ring->buf_pool && ring->buf_pool->rx_skb)
|
|
+ devm_kfree(dev, ring->buf_pool->rx_skb);
|
|
+ xgene_enet_free_desc_ring(ring->buf_pool);
|
|
+ xgene_enet_free_desc_ring(ring);
|
|
+}
|
|
+
|
|
+static struct xgene_enet_desc_ring *xgene_enet_create_desc_ring(
|
|
+ struct net_device *ndev, u32 ring_num,
|
|
+ enum xgene_enet_ring_cfgsize cfgsize, u32 ring_id)
|
|
+{
|
|
+ struct xgene_enet_desc_ring *ring;
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device *dev = ndev_to_dev(ndev);
|
|
+ u32 size;
|
|
+
|
|
+ ring = devm_kzalloc(dev, sizeof(struct xgene_enet_desc_ring),
|
|
+ GFP_KERNEL);
|
|
+ if (!ring)
|
|
+ return NULL;
|
|
+
|
|
+ ring->ndev = ndev;
|
|
+ ring->num = ring_num;
|
|
+ ring->cfgsize = cfgsize;
|
|
+ ring->id = ring_id;
|
|
+
|
|
+ size = xgene_enet_get_ring_size(dev, cfgsize);
|
|
+ ring->desc_addr = dma_zalloc_coherent(dev, size, &ring->dma,
|
|
+ GFP_KERNEL);
|
|
+ if (!ring->desc_addr) {
|
|
+ devm_kfree(dev, ring);
|
|
+ return NULL;
|
|
+ }
|
|
+ ring->size = size;
|
|
+
|
|
+ ring->cmd_base = pdata->ring_cmd_addr + (ring->num << 6);
|
|
+ ring->cmd = ring->cmd_base + INC_DEC_CMD_ADDR;
|
|
+ pdata->rm = RM3;
|
|
+ ring = xgene_enet_setup_ring(ring);
|
|
+ netdev_dbg(ndev, "ring info: num=%d size=%d id=%d slots=%d\n",
|
|
+ ring->num, ring->size, ring->id, ring->slots);
|
|
+
|
|
+ return ring;
|
|
+}
|
|
+
|
|
+static u16 xgene_enet_get_ring_id(enum xgene_ring_owner owner, u8 bufnum)
|
|
+{
|
|
+ return (owner << 6) | (bufnum & GENMASK(5, 0));
|
|
+}
|
|
+
|
|
+static int xgene_enet_create_desc_rings(struct net_device *ndev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct device *dev = ndev_to_dev(ndev);
|
|
+ struct xgene_enet_desc_ring *rx_ring, *tx_ring, *cp_ring;
|
|
+ struct xgene_enet_desc_ring *buf_pool = NULL;
|
|
+ u8 cpu_bufnum = 0, eth_bufnum = 0;
|
|
+ u8 bp_bufnum = 0x20;
|
|
+ u16 ring_id, ring_num = 0;
|
|
+ int ret;
|
|
+
|
|
+ /* allocate rx descriptor ring */
|
|
+ ring_id = xgene_enet_get_ring_id(RING_OWNER_CPU, cpu_bufnum++);
|
|
+ rx_ring = xgene_enet_create_desc_ring(ndev, ring_num++,
|
|
+ RING_CFGSIZE_16KB, ring_id);
|
|
+ if (!rx_ring) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ /* allocate buffer pool for receiving packets */
|
|
+ ring_id = xgene_enet_get_ring_id(RING_OWNER_ETH0, bp_bufnum++);
|
|
+ buf_pool = xgene_enet_create_desc_ring(ndev, ring_num++,
|
|
+ RING_CFGSIZE_2KB, ring_id);
|
|
+ if (!buf_pool) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ rx_ring->nbufpool = NUM_BUFPOOL;
|
|
+ rx_ring->buf_pool = buf_pool;
|
|
+ rx_ring->irq = pdata->rx_irq;
|
|
+ buf_pool->rx_skb = devm_kcalloc(dev, buf_pool->slots,
|
|
+ sizeof(struct sk_buff *), GFP_KERNEL);
|
|
+ if (!buf_pool->rx_skb) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ buf_pool->dst_ring_num = xgene_enet_dst_ring_num(buf_pool);
|
|
+ rx_ring->buf_pool = buf_pool;
|
|
+ pdata->rx_ring = rx_ring;
|
|
+
|
|
+ /* allocate tx descriptor ring */
|
|
+ ring_id = xgene_enet_get_ring_id(RING_OWNER_ETH0, eth_bufnum++);
|
|
+ tx_ring = xgene_enet_create_desc_ring(ndev, ring_num++,
|
|
+ RING_CFGSIZE_16KB, ring_id);
|
|
+ if (!tx_ring) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+ pdata->tx_ring = tx_ring;
|
|
+
|
|
+ cp_ring = pdata->rx_ring;
|
|
+ cp_ring->cp_skb = devm_kcalloc(dev, tx_ring->slots,
|
|
+ sizeof(struct sk_buff *), GFP_KERNEL);
|
|
+ if (!cp_ring->cp_skb) {
|
|
+ ret = -ENOMEM;
|
|
+ goto err;
|
|
+ }
|
|
+ pdata->tx_ring->cp_ring = cp_ring;
|
|
+ pdata->tx_ring->dst_ring_num = xgene_enet_dst_ring_num(cp_ring);
|
|
+
|
|
+ pdata->tx_qcnt_hi = pdata->tx_ring->slots / 2;
|
|
+ pdata->cp_qcnt_hi = pdata->rx_ring->slots / 2;
|
|
+ pdata->cp_qcnt_low = pdata->cp_qcnt_hi / 2;
|
|
+
|
|
+ return 0;
|
|
+
|
|
+err:
|
|
+ xgene_enet_free_desc_rings(pdata);
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static struct rtnl_link_stats64 *xgene_enet_get_stats64(
|
|
+ struct net_device *ndev,
|
|
+ struct rtnl_link_stats64 *storage)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ struct rtnl_link_stats64 *stats = &pdata->stats;
|
|
+
|
|
+ spin_lock(&pdata->stats_lock);
|
|
+ stats->rx_errors += stats->rx_length_errors +
|
|
+ stats->rx_crc_errors +
|
|
+ stats->rx_frame_errors +
|
|
+ stats->rx_fifo_errors;
|
|
+ memcpy(storage, &pdata->stats, sizeof(struct rtnl_link_stats64));
|
|
+ spin_unlock(&pdata->stats_lock);
|
|
+
|
|
+ return storage;
|
|
+}
|
|
+
|
|
+static int xgene_enet_set_mac_address(struct net_device *ndev, void *addr)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata = netdev_priv(ndev);
|
|
+ int ret;
|
|
+
|
|
+ ret = eth_mac_addr(ndev, addr);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+ xgene_gmac_set_mac_addr(pdata);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static const struct net_device_ops xgene_ndev_ops = {
|
|
+ .ndo_open = xgene_enet_open,
|
|
+ .ndo_stop = xgene_enet_close,
|
|
+ .ndo_start_xmit = xgene_enet_start_xmit,
|
|
+ .ndo_tx_timeout = xgene_enet_timeout,
|
|
+ .ndo_get_stats64 = xgene_enet_get_stats64,
|
|
+ .ndo_change_mtu = eth_change_mtu,
|
|
+ .ndo_set_mac_address = xgene_enet_set_mac_address,
|
|
+};
|
|
+
|
|
+static int xgene_enet_get_resources(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct platform_device *pdev;
|
|
+ struct net_device *ndev;
|
|
+ struct device *dev;
|
|
+ struct resource *res;
|
|
+ void *base_addr;
|
|
+ const char *mac;
|
|
+ int ret;
|
|
+
|
|
+ pdev = pdata->pdev;
|
|
+ dev = &pdev->dev;
|
|
+ ndev = pdata->ndev;
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "enet_csr");
|
|
+ if (!res) {
|
|
+ dev_err(dev, "Resource enet_csr not defined\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+ pdata->base_addr = devm_ioremap_resource(dev, res);
|
|
+ if (IS_ERR(pdata->base_addr)) {
|
|
+ dev_err(dev, "Unable to retrieve ENET Port CSR region\n");
|
|
+ return PTR_ERR(pdata->base_addr);
|
|
+ }
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ring_csr");
|
|
+ if (!res) {
|
|
+ dev_err(dev, "Resource ring_csr not defined\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+ pdata->ring_csr_addr = devm_ioremap_resource(dev, res);
|
|
+ if (IS_ERR(pdata->ring_csr_addr)) {
|
|
+ dev_err(dev, "Unable to retrieve ENET Ring CSR region\n");
|
|
+ return PTR_ERR(pdata->ring_csr_addr);
|
|
+ }
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ring_cmd");
|
|
+ if (!res) {
|
|
+ dev_err(dev, "Resource ring_cmd not defined\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+ pdata->ring_cmd_addr = devm_ioremap_resource(dev, res);
|
|
+ if (IS_ERR(pdata->ring_cmd_addr)) {
|
|
+ dev_err(dev, "Unable to retrieve ENET Ring command region\n");
|
|
+ return PTR_ERR(pdata->ring_cmd_addr);
|
|
+ }
|
|
+
|
|
+ ret = platform_get_irq(pdev, 0);
|
|
+ if (ret <= 0) {
|
|
+ dev_err(dev, "Unable to get ENET Rx IRQ\n");
|
|
+ ret = ret ? : -ENXIO;
|
|
+ return ret;
|
|
+ }
|
|
+ pdata->rx_irq = ret;
|
|
+
|
|
+ mac = of_get_mac_address(dev->of_node);
|
|
+ if (mac)
|
|
+ memcpy(ndev->dev_addr, mac, ndev->addr_len);
|
|
+ else
|
|
+ eth_hw_addr_random(ndev);
|
|
+ memcpy(ndev->perm_addr, ndev->dev_addr, ndev->addr_len);
|
|
+
|
|
+ pdata->phy_mode = of_get_phy_mode(pdev->dev.of_node);
|
|
+ if (pdata->phy_mode < 0) {
|
|
+ dev_err(dev, "Incorrect phy-connection-type in DTS\n");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ pdata->clk = devm_clk_get(&pdev->dev, NULL);
|
|
+ ret = IS_ERR(pdata->clk);
|
|
+ if (IS_ERR(pdata->clk)) {
|
|
+ dev_err(&pdev->dev, "can't get clock\n");
|
|
+ ret = PTR_ERR(pdata->clk);
|
|
+ return ret;
|
|
+ }
|
|
+
|
|
+ base_addr = pdata->base_addr;
|
|
+ pdata->eth_csr_addr = base_addr + BLOCK_ETH_CSR_OFFSET;
|
|
+ pdata->eth_ring_if_addr = base_addr + BLOCK_ETH_RING_IF_OFFSET;
|
|
+ pdata->eth_diag_csr_addr = base_addr + BLOCK_ETH_DIAG_CSR_OFFSET;
|
|
+ pdata->mcx_mac_addr = base_addr + BLOCK_ETH_MAC_OFFSET;
|
|
+ pdata->mcx_stats_addr = base_addr + BLOCK_ETH_STATS_OFFSET;
|
|
+ pdata->mcx_mac_csr_addr = base_addr + BLOCK_ETH_MAC_CSR_OFFSET;
|
|
+ pdata->rx_buff_cnt = NUM_PKT_BUF;
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_init_hw(struct xgene_enet_pdata *pdata)
|
|
+{
|
|
+ struct net_device *ndev = pdata->ndev;
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+ u16 dst_ring_num;
|
|
+ int ret;
|
|
+
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+
|
|
+ ret = xgene_enet_create_desc_rings(ndev);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "Error in ring configuration\n");
|
|
+ return ret;
|
|
+ }
|
|
+
|
|
+ /* setup buffer pool */
|
|
+ buf_pool = pdata->rx_ring->buf_pool;
|
|
+ xgene_enet_init_bufpool(buf_pool);
|
|
+ ret = xgene_enet_refill_bufpool(buf_pool, pdata->rx_buff_cnt);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ dst_ring_num = xgene_enet_dst_ring_num(pdata->rx_ring);
|
|
+ xgene_enet_cle_bypass(pdata, dst_ring_num, buf_pool->id);
|
|
+
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_probe(struct platform_device *pdev)
|
|
+{
|
|
+ struct net_device *ndev;
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct device *dev = &pdev->dev;
|
|
+ struct napi_struct *napi;
|
|
+ int ret;
|
|
+
|
|
+ ndev = alloc_etherdev(sizeof(struct xgene_enet_pdata));
|
|
+ if (!ndev)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ pdata = netdev_priv(ndev);
|
|
+
|
|
+ pdata->pdev = pdev;
|
|
+ pdata->ndev = ndev;
|
|
+ SET_NETDEV_DEV(ndev, dev);
|
|
+ platform_set_drvdata(pdev, pdata);
|
|
+ ndev->netdev_ops = &xgene_ndev_ops;
|
|
+ xgene_enet_set_ethtool_ops(ndev);
|
|
+ ndev->features |= NETIF_F_IP_CSUM |
|
|
+ NETIF_F_GSO |
|
|
+ NETIF_F_GRO;
|
|
+
|
|
+ ret = xgene_enet_get_resources(pdata);
|
|
+ if (ret)
|
|
+ goto err;
|
|
+
|
|
+ xgene_enet_reset(pdata);
|
|
+ xgene_gmac_init(pdata, SPEED_1000);
|
|
+
|
|
+ spin_lock_init(&pdata->stats_lock);
|
|
+ ret = register_netdev(ndev);
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "Failed to register netdev\n");
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(64));
|
|
+ if (ret) {
|
|
+ netdev_err(ndev, "No usable DMA configuration\n");
|
|
+ goto err;
|
|
+ }
|
|
+
|
|
+ ret = xgene_enet_init_hw(pdata);
|
|
+ if (ret)
|
|
+ goto err;
|
|
+
|
|
+ napi = &pdata->rx_ring->napi;
|
|
+ netif_napi_add(ndev, napi, xgene_enet_napi, NAPI_POLL_WEIGHT);
|
|
+ ret = xgene_enet_mdio_config(pdata);
|
|
+
|
|
+ return ret;
|
|
+err:
|
|
+ free_netdev(ndev);
|
|
+ return ret;
|
|
+}
|
|
+
|
|
+static int xgene_enet_remove(struct platform_device *pdev)
|
|
+{
|
|
+ struct xgene_enet_pdata *pdata;
|
|
+ struct net_device *ndev;
|
|
+
|
|
+ pdata = platform_get_drvdata(pdev);
|
|
+ ndev = pdata->ndev;
|
|
+
|
|
+ xgene_gmac_rx_disable(pdata);
|
|
+ xgene_gmac_tx_disable(pdata);
|
|
+
|
|
+ netif_napi_del(&pdata->rx_ring->napi);
|
|
+ xgene_enet_mdio_remove(pdata);
|
|
+ xgene_enet_delete_desc_rings(pdata);
|
|
+ unregister_netdev(ndev);
|
|
+ xgene_gport_shutdown(pdata);
|
|
+ free_netdev(ndev);
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static struct of_device_id xgene_enet_match[] = {
|
|
+ {.compatible = "apm,xgene-enet",},
|
|
+ {},
|
|
+};
|
|
+
|
|
+MODULE_DEVICE_TABLE(of, xgene_enet_match);
|
|
+
|
|
+static struct platform_driver xgene_enet_driver = {
|
|
+ .driver = {
|
|
+ .name = "xgene-enet",
|
|
+ .owner = THIS_MODULE,
|
|
+ .of_match_table = xgene_enet_match,
|
|
+ },
|
|
+ .probe = xgene_enet_probe,
|
|
+ .remove = xgene_enet_remove,
|
|
+};
|
|
+
|
|
+module_platform_driver(xgene_enet_driver);
|
|
+
|
|
+MODULE_DESCRIPTION("APM X-Gene SoC Ethernet driver");
|
|
+MODULE_VERSION(XGENE_DRV_VERSION);
|
|
+MODULE_AUTHOR("Keyur Chudgar <kchudgar@apm.com>");
|
|
+MODULE_LICENSE("GPL");
|
|
diff --git a/drivers/net/ethernet/apm/xgene/xgene_enet_main.h b/drivers/net/ethernet/apm/xgene/xgene_enet_main.h
|
|
new file mode 100644
|
|
index 0000000..f4f7e4a
|
|
--- /dev/null
|
|
+++ b/drivers/net/ethernet/apm/xgene/xgene_enet_main.h
|
|
@@ -0,0 +1,107 @@
|
|
+/* Applied Micro X-Gene SoC Ethernet Driver
|
|
+ *
|
|
+ * Copyright (c) 2014, Applied Micro Circuits Corporation
|
|
+ * Authors: Iyappan Subramanian <isubramanian@apm.com>
|
|
+ * Ravi Patel <rapatel@apm.com>
|
|
+ * Keyur Chudgar <kchudgar@apm.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ * You should have received a copy of the GNU General Public License
|
|
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
+ */
|
|
+
|
|
+#ifndef __XGENE_ENET_MAIN_H__
|
|
+#define __XGENE_ENET_MAIN_H__
|
|
+
|
|
+#include <linux/clk.h>
|
|
+#include <linux/of_platform.h>
|
|
+#include <linux/of_net.h>
|
|
+#include <linux/of_mdio.h>
|
|
+#include <linux/module.h>
|
|
+#include <net/ip.h>
|
|
+#include <linux/prefetch.h>
|
|
+#include <linux/if_vlan.h>
|
|
+#include <linux/phy.h>
|
|
+#include "xgene_enet_hw.h"
|
|
+
|
|
+#define XGENE_DRV_VERSION "v1.0"
|
|
+#define XGENE_ENET_MAX_MTU 1536
|
|
+#define SKB_BUFFER_SIZE (XGENE_ENET_MAX_MTU - NET_IP_ALIGN)
|
|
+#define NUM_PKT_BUF 64
|
|
+#define NUM_BUFPOOL 32
|
|
+
|
|
+/* software context of a descriptor ring */
|
|
+struct xgene_enet_desc_ring {
|
|
+ struct net_device *ndev;
|
|
+ u16 id;
|
|
+ u16 num;
|
|
+ u16 head;
|
|
+ u16 tail;
|
|
+ u16 slots;
|
|
+ u16 irq;
|
|
+ u32 size;
|
|
+ u32 state[NUM_RING_CONFIG];
|
|
+ void __iomem *cmd_base;
|
|
+ void __iomem *cmd;
|
|
+ dma_addr_t dma;
|
|
+ u16 dst_ring_num;
|
|
+ u8 nbufpool;
|
|
+ struct sk_buff *(*rx_skb);
|
|
+ struct sk_buff *(*cp_skb);
|
|
+ enum xgene_enet_ring_cfgsize cfgsize;
|
|
+ struct xgene_enet_desc_ring *cp_ring;
|
|
+ struct xgene_enet_desc_ring *buf_pool;
|
|
+ struct napi_struct napi;
|
|
+ union {
|
|
+ void *desc_addr;
|
|
+ struct xgene_enet_raw_desc *raw_desc;
|
|
+ struct xgene_enet_raw_desc16 *raw_desc16;
|
|
+ };
|
|
+};
|
|
+
|
|
+/* ethernet private data */
|
|
+struct xgene_enet_pdata {
|
|
+ struct net_device *ndev;
|
|
+ struct mii_bus *mdio_bus;
|
|
+ struct phy_device *phy_dev;
|
|
+ int phy_speed;
|
|
+ struct clk *clk;
|
|
+ struct platform_device *pdev;
|
|
+ struct xgene_enet_desc_ring *tx_ring;
|
|
+ struct xgene_enet_desc_ring *rx_ring;
|
|
+ char *dev_name;
|
|
+ u32 rx_buff_cnt;
|
|
+ u32 tx_qcnt_hi;
|
|
+ u32 cp_qcnt_hi;
|
|
+ u32 cp_qcnt_low;
|
|
+ u32 rx_irq;
|
|
+ void __iomem *eth_csr_addr;
|
|
+ void __iomem *eth_ring_if_addr;
|
|
+ void __iomem *eth_diag_csr_addr;
|
|
+ void __iomem *mcx_mac_addr;
|
|
+ void __iomem *mcx_stats_addr;
|
|
+ void __iomem *mcx_mac_csr_addr;
|
|
+ void __iomem *base_addr;
|
|
+ void __iomem *ring_csr_addr;
|
|
+ void __iomem *ring_cmd_addr;
|
|
+ u32 phy_addr;
|
|
+ int phy_mode;
|
|
+ u32 speed;
|
|
+ u16 rm;
|
|
+ struct rtnl_link_stats64 stats;
|
|
+ /* statistics lock */
|
|
+ spinlock_t stats_lock;
|
|
+};
|
|
+
|
|
+void xgene_enet_set_ethtool_ops(struct net_device *netdev);
|
|
+
|
|
+#endif /* __XGENE_ENET_MAIN_H__ */
|
|
diff --git a/drivers/of/address.c b/drivers/of/address.c
|
|
index e371825..5eaadae 100644
|
|
--- a/drivers/of/address.c
|
|
+++ b/drivers/of/address.c
|
|
@@ -5,6 +5,7 @@
|
|
#include <linux/module.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/pci_regs.h>
|
|
+#include <linux/slab.h>
|
|
#include <linux/string.h>
|
|
|
|
/* Max address size we deal with */
|
|
@@ -601,12 +602,72 @@ const __be32 *of_get_address(struct device_node *dev, int index, u64 *size,
|
|
}
|
|
EXPORT_SYMBOL(of_get_address);
|
|
|
|
+struct io_range {
|
|
+ struct list_head list;
|
|
+ phys_addr_t start;
|
|
+ resource_size_t size;
|
|
+};
|
|
+
|
|
+static LIST_HEAD(io_range_list);
|
|
+
|
|
+/*
|
|
+ * Record the PCI IO range (expressed as CPU physical address + size).
|
|
+ * Return a negative value if an error has occured, zero otherwise
|
|
+ */
|
|
+int __weak pci_register_io_range(phys_addr_t addr, resource_size_t size)
|
|
+{
|
|
+#ifdef PCI_IOBASE
|
|
+ struct io_range *res;
|
|
+ resource_size_t allocated_size = 0;
|
|
+
|
|
+ /* check if the range hasn't been previously recorded */
|
|
+ list_for_each_entry(res, &io_range_list, list) {
|
|
+ if (addr >= res->start && addr + size <= res->start + size)
|
|
+ return 0;
|
|
+ allocated_size += res->size;
|
|
+ }
|
|
+
|
|
+ /* range not registed yet, check for available space */
|
|
+ if (allocated_size + size - 1 > IO_SPACE_LIMIT)
|
|
+ return -E2BIG;
|
|
+
|
|
+ /* add the range to the list */
|
|
+ res = kzalloc(sizeof(*res), GFP_KERNEL);
|
|
+ if (!res)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ res->start = addr;
|
|
+ res->size = size;
|
|
+
|
|
+ list_add_tail(&res->list, &io_range_list);
|
|
+
|
|
+ return 0;
|
|
+#else
|
|
+ return -EINVAL;
|
|
+#endif
|
|
+}
|
|
+
|
|
unsigned long __weak pci_address_to_pio(phys_addr_t address)
|
|
{
|
|
+#ifdef PCI_IOBASE
|
|
+ struct io_range *res;
|
|
+ resource_size_t offset = 0;
|
|
+
|
|
+ list_for_each_entry(res, &io_range_list, list) {
|
|
+ if (address >= res->start &&
|
|
+ address < res->start + res->size) {
|
|
+ return res->start - address + offset;
|
|
+ }
|
|
+ offset += res->size;
|
|
+ }
|
|
+
|
|
+ return (unsigned long)-1;
|
|
+#else
|
|
if (address > IO_SPACE_LIMIT)
|
|
return (unsigned long)-1;
|
|
|
|
return (unsigned long) address;
|
|
+#endif
|
|
}
|
|
|
|
static int __of_address_to_resource(struct device_node *dev,
|
|
@@ -847,3 +908,50 @@ bool of_dma_is_coherent(struct device_node *np)
|
|
return false;
|
|
}
|
|
EXPORT_SYMBOL_GPL(of_dma_is_coherent);
|
|
+
|
|
+/*
|
|
+ * of_pci_range_to_resource - Create a resource from an of_pci_range
|
|
+ * @range: the PCI range that describes the resource
|
|
+ * @np: device node where the range belongs to
|
|
+ * @res: pointer to a valid resource that will be updated to
|
|
+ * reflect the values contained in the range.
|
|
+ *
|
|
+ * Returns EINVAL if the range cannot be converted to resource.
|
|
+ *
|
|
+ * Note that if the range is an IO range, the resource will be converted
|
|
+ * using pci_address_to_pio() which can fail if it is called too early or
|
|
+ * if the range cannot be matched to any host bridge IO space (our case here).
|
|
+ * To guard against that we try to register the IO range first.
|
|
+ * If that fails we know that pci_address_to_pio() will do too.
|
|
+ */
|
|
+int of_pci_range_to_resource(struct of_pci_range *range,
|
|
+ struct device_node *np, struct resource *res)
|
|
+{
|
|
+ int err;
|
|
+ res->flags = range->flags;
|
|
+ res->parent = res->child = res->sibling = NULL;
|
|
+ res->name = np->full_name;
|
|
+
|
|
+ if (res->flags & IORESOURCE_IO) {
|
|
+ unsigned long port = -1;
|
|
+ err = pci_register_io_range(range->cpu_addr, range->size);
|
|
+ if (err)
|
|
+ goto invalid_range;
|
|
+ port = pci_address_to_pio(range->cpu_addr);
|
|
+ if (port == (unsigned long)-1) {
|
|
+ err = -EINVAL;
|
|
+ goto invalid_range;
|
|
+ }
|
|
+ res->start = port;
|
|
+ } else {
|
|
+ res->start = range->cpu_addr;
|
|
+ }
|
|
+ res->end = res->start + range->size - 1;
|
|
+ return 0;
|
|
+
|
|
+invalid_range:
|
|
+ res->start = (resource_size_t)OF_BAD_ADDR;
|
|
+ res->end = (resource_size_t)OF_BAD_ADDR;
|
|
+ return err;
|
|
+}
|
|
+
|
|
diff --git a/drivers/of/of_pci.c b/drivers/of/of_pci.c
|
|
index 8481996..e81402a 100644
|
|
--- a/drivers/of/of_pci.c
|
|
+++ b/drivers/of/of_pci.c
|
|
@@ -1,6 +1,7 @@
|
|
#include <linux/kernel.h>
|
|
#include <linux/export.h>
|
|
#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
#include <linux/of_pci.h>
|
|
|
|
static inline int __of_pci_pci_compare(struct device_node *node,
|
|
@@ -89,6 +90,141 @@ int of_pci_parse_bus_range(struct device_node *node, struct resource *res)
|
|
}
|
|
EXPORT_SYMBOL_GPL(of_pci_parse_bus_range);
|
|
|
|
+/**
|
|
+ * pci_host_bridge_of_get_ranges - Parse PCI host bridge resources from DT
|
|
+ * @dev: device node of the host bridge having the range property
|
|
+ * @resources: list where the range of resources will be added after DT parsing
|
|
+ * @io_base: pointer to a variable that will contain the physical address for
|
|
+ * the start of the I/O range.
|
|
+ *
|
|
+ * It is the callers job to free the @resources list if an error is returned.
|
|
+ *
|
|
+ * This function will parse the "ranges" property of a PCI host bridge device
|
|
+ * node and setup the resource mapping based on its content. It is expected
|
|
+ * that the property conforms with the Power ePAPR document.
|
|
+ *
|
|
+ * Each architecture is then offered the chance of applying their own
|
|
+ * filtering of pci_host_bridge_windows based on their own restrictions by
|
|
+ * calling pcibios_fixup_bridge_ranges(). The filtered list of windows
|
|
+ * can then be used when creating a pci_host_bridge structure.
|
|
+ */
|
|
+static int pci_host_bridge_of_get_ranges(struct device_node *dev,
|
|
+ struct list_head *resources, resource_size_t *io_base)
|
|
+{
|
|
+ struct resource *res;
|
|
+ struct of_pci_range range;
|
|
+ struct of_pci_range_parser parser;
|
|
+ int err;
|
|
+
|
|
+ pr_info("PCI host bridge %s ranges:\n", dev->full_name);
|
|
+
|
|
+ /* Check for ranges property */
|
|
+ err = of_pci_range_parser_init(&parser, dev);
|
|
+ if (err)
|
|
+ return err;
|
|
+
|
|
+ pr_debug("Parsing ranges property...\n");
|
|
+ for_each_of_pci_range(&parser, &range) {
|
|
+ /* Read next ranges element */
|
|
+ pr_debug("pci_space: 0x%08x pci_addr:0x%016llx cpu_addr:0x%016llx size:0x%016llx\n",
|
|
+ range.pci_space, range.pci_addr, range.cpu_addr, range.size);
|
|
+
|
|
+ /*
|
|
+ * If we failed translation or got a zero-sized region
|
|
+ * then skip this range
|
|
+ */
|
|
+ if (range.cpu_addr == OF_BAD_ADDR || range.size == 0)
|
|
+ continue;
|
|
+
|
|
+ res = kzalloc(sizeof(struct resource), GFP_KERNEL);
|
|
+ if (!res)
|
|
+ return -ENOMEM;
|
|
+
|
|
+ err = of_pci_range_to_resource(&range, dev, res);
|
|
+ if (err)
|
|
+ return err;
|
|
+
|
|
+ if (resource_type(res) == IORESOURCE_IO)
|
|
+ *io_base = range.cpu_addr;
|
|
+
|
|
+ pci_add_resource_offset(resources, res,
|
|
+ res->start - range.pci_addr);
|
|
+ }
|
|
+
|
|
+ /* Apply architecture specific fixups for the ranges */
|
|
+ return pcibios_fixup_bridge_ranges(resources);
|
|
+}
|
|
+
|
|
+static atomic_t domain_nr = ATOMIC_INIT(-1);
|
|
+
|
|
+/**
|
|
+ * of_create_pci_host_bridge - Create a PCI host bridge structure using
|
|
+ * information passed in the DT.
|
|
+ * @parent: device owning this host bridge
|
|
+ * @ops: pci_ops associated with the host controller
|
|
+ * @host_data: opaque data structure used by the host controller.
|
|
+ *
|
|
+ * returns a pointer to the newly created pci_host_bridge structure, or
|
|
+ * NULL if the call failed.
|
|
+ *
|
|
+ * This function will try to obtain the host bridge domain number by
|
|
+ * using of_alias_get_id() call with "pci-domain" as a stem. If that
|
|
+ * fails, a local allocator will be used that will put each host bridge
|
|
+ * in a new domain.
|
|
+ */
|
|
+struct pci_host_bridge *
|
|
+of_create_pci_host_bridge(struct device *parent, struct pci_ops *ops, void *host_data)
|
|
+{
|
|
+ int err, domain, busno;
|
|
+ struct resource *bus_range;
|
|
+ struct pci_bus *root_bus;
|
|
+ struct pci_host_bridge *bridge;
|
|
+ resource_size_t io_base = 0;
|
|
+ LIST_HEAD(res);
|
|
+
|
|
+ bus_range = kzalloc(sizeof(*bus_range), GFP_KERNEL);
|
|
+ if (!bus_range)
|
|
+ return ERR_PTR(-ENOMEM);
|
|
+
|
|
+ domain = of_alias_get_id(parent->of_node, "pci-domain");
|
|
+ if (domain == -ENODEV)
|
|
+ domain = atomic_inc_return(&domain_nr);
|
|
+
|
|
+ err = of_pci_parse_bus_range(parent->of_node, bus_range);
|
|
+ if (err) {
|
|
+ dev_info(parent, "No bus range for %s, using default [0-255]\n",
|
|
+ parent->of_node->full_name);
|
|
+ bus_range->start = 0;
|
|
+ bus_range->end = 255;
|
|
+ bus_range->flags = IORESOURCE_BUS;
|
|
+ }
|
|
+ busno = bus_range->start;
|
|
+ pci_add_resource(&res, bus_range);
|
|
+
|
|
+ /* now parse the rest of host bridge bus ranges */
|
|
+ err = pci_host_bridge_of_get_ranges(parent->of_node, &res, &io_base);
|
|
+ if (err)
|
|
+ goto err_create;
|
|
+
|
|
+ /* then create the root bus */
|
|
+ root_bus = pci_create_root_bus_in_domain(parent, domain, busno,
|
|
+ ops, host_data, &res);
|
|
+ if (IS_ERR(root_bus)) {
|
|
+ err = PTR_ERR(root_bus);
|
|
+ goto err_create;
|
|
+ }
|
|
+
|
|
+ bridge = to_pci_host_bridge(root_bus->bridge);
|
|
+ bridge->io_base = io_base;
|
|
+
|
|
+ return bridge;
|
|
+
|
|
+err_create:
|
|
+ pci_free_resource_list(&res);
|
|
+ return ERR_PTR(err);
|
|
+}
|
|
+EXPORT_SYMBOL_GPL(of_create_pci_host_bridge);
|
|
+
|
|
#ifdef CONFIG_PCI_MSI
|
|
|
|
static LIST_HEAD(of_pci_msi_chip_list);
|
|
diff --git a/drivers/pci/host-bridge.c b/drivers/pci/host-bridge.c
|
|
index 0e5f3c9..54ceafd 100644
|
|
--- a/drivers/pci/host-bridge.c
|
|
+++ b/drivers/pci/host-bridge.c
|
|
@@ -16,12 +16,13 @@ static struct pci_bus *find_pci_root_bus(struct pci_bus *bus)
|
|
return bus;
|
|
}
|
|
|
|
-static struct pci_host_bridge *find_pci_host_bridge(struct pci_bus *bus)
|
|
+struct pci_host_bridge *find_pci_host_bridge(struct pci_bus *bus)
|
|
{
|
|
struct pci_bus *root_bus = find_pci_root_bus(bus);
|
|
|
|
return to_pci_host_bridge(root_bus->bridge);
|
|
}
|
|
+EXPORT_SYMBOL_GPL(find_pci_host_bridge);
|
|
|
|
void pci_set_host_bridge_release(struct pci_host_bridge *bridge,
|
|
void (*release_fn)(struct pci_host_bridge *),
|
|
@@ -82,3 +83,18 @@ void pcibios_bus_to_resource(struct pci_bus *bus, struct resource *res,
|
|
res->end = region->end + offset;
|
|
}
|
|
EXPORT_SYMBOL(pcibios_bus_to_resource);
|
|
+
|
|
+/**
|
|
+ * Simple version of the platform specific code for filtering the list
|
|
+ * of resources obtained from the ranges declaration in DT.
|
|
+ *
|
|
+ * Platforms can override this function in order to impose stronger
|
|
+ * constraints onto the list of resources that a host bridge can use.
|
|
+ * The filtered list will then be used to create a root bus and associate
|
|
+ * it with the host bridge.
|
|
+ *
|
|
+ */
|
|
+int __weak pcibios_fixup_bridge_ranges(struct list_head *resources)
|
|
+{
|
|
+ return 0;
|
|
+}
|
|
diff --git a/drivers/pci/host/Kconfig b/drivers/pci/host/Kconfig
|
|
index 2d8a4d0..76daa18 100644
|
|
--- a/drivers/pci/host/Kconfig
|
|
+++ b/drivers/pci/host/Kconfig
|
|
@@ -46,6 +46,16 @@ config PCI_HOST_GENERIC
|
|
Say Y here if you want to support a simple generic PCI host
|
|
controller, such as the one emulated by kvmtool.
|
|
|
|
+config PCI_XGENE
|
|
+ bool "X-Gene PCIe controller"
|
|
+ depends on ARCH_XGENE
|
|
+ depends on OF
|
|
+ select PCIEPORTBUS
|
|
+ help
|
|
+ Say Y here if you want internal PCI support on APM X-Gene SoC.
|
|
+ There are 5 internal PCIe ports available. Each port is GEN3 capable
|
|
+ and have varied lanes from x1 to x8.
|
|
+
|
|
config PCIE_SPEAR13XX
|
|
tristate "STMicroelectronics SPEAr PCIe controller"
|
|
depends on ARCH_SPEAR13XX
|
|
diff --git a/drivers/pci/host/Makefile b/drivers/pci/host/Makefile
|
|
index 0daec79..6990412 100644
|
|
--- a/drivers/pci/host/Makefile
|
|
+++ b/drivers/pci/host/Makefile
|
|
@@ -6,4 +6,5 @@ obj-$(CONFIG_PCI_TEGRA) += pci-tegra.o
|
|
obj-$(CONFIG_PCI_RCAR_GEN2) += pci-rcar-gen2.o
|
|
obj-$(CONFIG_PCI_RCAR_GEN2_PCIE) += pcie-rcar.o
|
|
obj-$(CONFIG_PCI_HOST_GENERIC) += pci-host-generic.o
|
|
+obj-$(CONFIG_PCI_XGENE) += pci-xgene.o
|
|
obj-$(CONFIG_PCIE_SPEAR13XX) += pcie-spear13xx.o
|
|
diff --git a/drivers/pci/host/pci-xgene.c b/drivers/pci/host/pci-xgene.c
|
|
new file mode 100644
|
|
index 0000000..7bf4ac7
|
|
--- /dev/null
|
|
+++ b/drivers/pci/host/pci-xgene.c
|
|
@@ -0,0 +1,725 @@
|
|
+/**
|
|
+ * APM X-Gene PCIe Driver
|
|
+ *
|
|
+ * Copyright (c) 2013 Applied Micro Circuits Corporation.
|
|
+ *
|
|
+ * Author: Tanmay Inamdar <tinamdar@apm.com>.
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; either version 2 of the License, or (at your
|
|
+ * option) any later version.
|
|
+ *
|
|
+ * This program is distributed in the hope that it will be useful,
|
|
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
+ * GNU General Public License for more details.
|
|
+ *
|
|
+ */
|
|
+#include <linux/clk-private.h>
|
|
+#include <linux/delay.h>
|
|
+#include <linux/io.h>
|
|
+#include <linux/jiffies.h>
|
|
+#include <linux/memblock.h>
|
|
+#include <linux/module.h>
|
|
+#include <linux/of.h>
|
|
+#include <linux/of_address.h>
|
|
+#include <linux/of_irq.h>
|
|
+#include <linux/of_pci.h>
|
|
+#include <linux/pci.h>
|
|
+#include <linux/platform_device.h>
|
|
+#include <linux/slab.h>
|
|
+
|
|
+#define PCIECORE_LTSSM 0x4c
|
|
+#define PCIECORE_CTLANDSTATUS 0x50
|
|
+#define INTXSTATUSMASK 0x6c
|
|
+#define PIM1_1L 0x80
|
|
+#define IBAR2 0x98
|
|
+#define IR2MSK 0x9c
|
|
+#define PIM2_1L 0xa0
|
|
+#define IBAR3L 0xb4
|
|
+#define IR3MSKL 0xbc
|
|
+#define PIM3_1L 0xc4
|
|
+#define OMR1BARL 0x100
|
|
+#define OMR2BARL 0x118
|
|
+#define OMR3BARL 0x130
|
|
+#define CFGBARL 0x154
|
|
+#define CFGBARH 0x158
|
|
+#define CFGCTL 0x15c
|
|
+#define RTDID 0x160
|
|
+#define BRIDGE_CFG_0 0x2000
|
|
+#define BRIDGE_CFG_1 0x2004
|
|
+#define BRIDGE_CFG_4 0x2010
|
|
+#define BRIDGE_CFG_32 0x2030
|
|
+#define BRIDGE_CFG_14 0x2038
|
|
+#define BRIDGE_CTRL_1 0x2204
|
|
+#define BRIDGE_CTRL_2 0x2208
|
|
+#define BRIDGE_CTRL_5 0x2214
|
|
+#define BRIDGE_STATUS_0 0x2600
|
|
+#define MEM_RAM_SHUTDOWN 0xd070
|
|
+#define BLOCK_MEM_RDY 0xd074
|
|
+
|
|
+#define DEVICE_PORT_TYPE_MASK 0x03c00000
|
|
+#define PM_FORCE_RP_MODE_MASK 0x00000400
|
|
+#define SWITCH_PORT_MODE_MASK 0x00000800
|
|
+#define CLASS_CODE_MASK 0xffffff00
|
|
+#define LINK_UP_MASK 0x00000100
|
|
+#define AER_OPTIONAL_ERROR_EN 0xffc00000
|
|
+#define XGENE_PCIE_DEV_CTRL 0x2f0f
|
|
+#define AXI_EP_CFG_ACCESS 0x10000
|
|
+#define ENABLE_ASPM 0x08000000
|
|
+#define XGENE_PORT_TYPE_RC 0x05000000
|
|
+#define BLOCK_MEM_RDY_VAL 0xFFFFFFFF
|
|
+#define EN_COHERENCY 0xF0000000
|
|
+#define EN_REG 0x00000001
|
|
+#define OB_LO_IO 0x00000002
|
|
+#define XGENE_PCIE_VENDORID 0xE008
|
|
+#define XGENE_PCIE_DEVICEID 0xE004
|
|
+#define XGENE_PCIE_ECC_TIMEOUT 10 /* ms */
|
|
+#define XGENE_LTSSM_DETECT_WAIT 20 /* ms */
|
|
+#define XGENE_LTSSM_L0_WAIT 4 /* ms */
|
|
+#define SZ_1T (SZ_1G*1024ULL)
|
|
+#define PIPE_PHY_RATE_RD(src) ((0xc000 & (u32)(src)) >> 0xe)
|
|
+
|
|
+struct xgene_pcie_port {
|
|
+ struct device_node *node;
|
|
+ struct device *dev;
|
|
+ struct clk *clk;
|
|
+ void __iomem *csr_base;
|
|
+ void __iomem *cfg_base;
|
|
+ u8 link_up;
|
|
+};
|
|
+
|
|
+static inline u32 pcie_bar_low_val(u32 addr, u32 flags)
|
|
+{
|
|
+ return (addr & PCI_BASE_ADDRESS_MEM_MASK) | flags;
|
|
+}
|
|
+
|
|
+/* PCIE Configuration Out/In */
|
|
+static inline void xgene_pcie_cfg_out32(void __iomem *addr, int offset, u32 val)
|
|
+{
|
|
+ writel(val, addr + offset);
|
|
+}
|
|
+
|
|
+static inline void xgene_pcie_cfg_out16(void __iomem *addr, int offset, u16 val)
|
|
+{
|
|
+ u32 val32 = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 2:
|
|
+ val32 &= ~0xFFFF0000;
|
|
+ val32 |= (u32)val << 16;
|
|
+ break;
|
|
+ case 0:
|
|
+ default:
|
|
+ val32 &= ~0xFFFF;
|
|
+ val32 |= val;
|
|
+ break;
|
|
+ }
|
|
+ writel(val32, addr + (offset & ~0x3));
|
|
+}
|
|
+
|
|
+static inline void xgene_pcie_cfg_out8(void __iomem *addr, int offset, u8 val)
|
|
+{
|
|
+ u32 val32 = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 0:
|
|
+ val32 &= ~0xFF;
|
|
+ val32 |= val;
|
|
+ break;
|
|
+ case 1:
|
|
+ val32 &= ~0xFF00;
|
|
+ val32 |= (u32)val << 8;
|
|
+ break;
|
|
+ case 2:
|
|
+ val32 &= ~0xFF0000;
|
|
+ val32 |= (u32)val << 16;
|
|
+ break;
|
|
+ case 3:
|
|
+ default:
|
|
+ val32 &= ~0xFF000000;
|
|
+ val32 |= (u32)val << 24;
|
|
+ break;
|
|
+ }
|
|
+ writel(val32, addr + (offset & ~0x3));
|
|
+}
|
|
+
|
|
+static inline void xgene_pcie_cfg_in32(void __iomem *addr, int offset, u32 *val)
|
|
+{
|
|
+ *val = readl(addr + offset);
|
|
+}
|
|
+
|
|
+static inline void
|
|
+xgene_pcie_cfg_in16(void __iomem *addr, int offset, u32 *val)
|
|
+{
|
|
+ *val = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 2:
|
|
+ *val >>= 16;
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ *val &= 0xFFFF;
|
|
+}
|
|
+
|
|
+static inline void
|
|
+xgene_pcie_cfg_in8(void __iomem *addr, int offset, u32 *val)
|
|
+{
|
|
+ *val = readl(addr + (offset & ~0x3));
|
|
+
|
|
+ switch (offset & 0x3) {
|
|
+ case 3:
|
|
+ *val = *val >> 24;
|
|
+ break;
|
|
+ case 2:
|
|
+ *val = *val >> 16;
|
|
+ break;
|
|
+ case 1:
|
|
+ *val = *val >> 8;
|
|
+ break;
|
|
+ }
|
|
+ *val &= 0xFF;
|
|
+}
|
|
+
|
|
+/* When the address bit [17:16] is 2'b01, the Configuration access will be
|
|
+ * treated as Type 1 and it will be forwarded to external PCIe device.
|
|
+ */
|
|
+static void __iomem *xgene_pcie_get_cfg_base(struct pci_bus *bus)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+
|
|
+ if (bus->number >= (bus->primary + 1))
|
|
+ return port->cfg_base + AXI_EP_CFG_ACCESS;
|
|
+
|
|
+ return port->cfg_base;
|
|
+}
|
|
+
|
|
+/* For Configuration request, RTDID register is used as Bus Number,
|
|
+ * Device Number and Function number of the header fields.
|
|
+ */
|
|
+static void xgene_pcie_set_rtdid_reg(struct pci_bus *bus, uint devfn)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+ unsigned int b, d, f;
|
|
+ u32 rtdid_val = 0;
|
|
+
|
|
+ b = bus->number;
|
|
+ d = PCI_SLOT(devfn);
|
|
+ f = PCI_FUNC(devfn);
|
|
+
|
|
+ if (!pci_is_root_bus(bus))
|
|
+ rtdid_val = (b << 8) | (d << 3) | f;
|
|
+
|
|
+ writel(rtdid_val, port->csr_base + RTDID);
|
|
+ /* read the register back to ensure flush */
|
|
+ readl(port->csr_base + RTDID);
|
|
+}
|
|
+
|
|
+static int xgene_pcie_read_config(struct pci_bus *bus, unsigned int devfn,
|
|
+ int offset, int len, u32 *val)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+ void __iomem *addr;
|
|
+
|
|
+ if ((pci_is_root_bus(bus) && devfn != 0) || !port->link_up)
|
|
+ return PCIBIOS_DEVICE_NOT_FOUND;
|
|
+
|
|
+ xgene_pcie_set_rtdid_reg(bus, devfn);
|
|
+ addr = xgene_pcie_get_cfg_base(bus);
|
|
+ switch (len) {
|
|
+ case 1:
|
|
+ xgene_pcie_cfg_in8(addr, offset, val);
|
|
+ break;
|
|
+ case 2:
|
|
+ xgene_pcie_cfg_in16(addr, offset, val);
|
|
+ break;
|
|
+ default:
|
|
+ xgene_pcie_cfg_in32(addr, offset, val);
|
|
+ break;
|
|
+ }
|
|
+ return PCIBIOS_SUCCESSFUL;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_write_config(struct pci_bus *bus, unsigned int devfn,
|
|
+ int offset, int len, u32 val)
|
|
+{
|
|
+ struct xgene_pcie_port *port = bus->sysdata;
|
|
+ void __iomem *addr;
|
|
+
|
|
+ if ((pci_is_root_bus(bus) && devfn != 0) || !port->link_up)
|
|
+ return PCIBIOS_DEVICE_NOT_FOUND;
|
|
+
|
|
+ xgene_pcie_set_rtdid_reg(bus, devfn);
|
|
+ addr = xgene_pcie_get_cfg_base(bus);
|
|
+ switch (len) {
|
|
+ case 1:
|
|
+ xgene_pcie_cfg_out8(addr, offset, (u8)val);
|
|
+ break;
|
|
+ case 2:
|
|
+ xgene_pcie_cfg_out16(addr, offset, (u16)val);
|
|
+ break;
|
|
+ default:
|
|
+ xgene_pcie_cfg_out32(addr, offset, val);
|
|
+ break;
|
|
+ }
|
|
+ return PCIBIOS_SUCCESSFUL;
|
|
+}
|
|
+
|
|
+static struct pci_ops xgene_pcie_ops = {
|
|
+ .read = xgene_pcie_read_config,
|
|
+ .write = xgene_pcie_write_config
|
|
+};
|
|
+
|
|
+static void xgene_pcie_program_core(void __iomem *csr_base)
|
|
+{
|
|
+ u32 val;
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_0);
|
|
+ val |= AER_OPTIONAL_ERROR_EN;
|
|
+ writel(val, csr_base + BRIDGE_CFG_0);
|
|
+ writel(0x0, csr_base + INTXSTATUSMASK);
|
|
+ val = readl(csr_base + BRIDGE_CTRL_1);
|
|
+ val = (val & ~0xffff) | XGENE_PCIE_DEV_CTRL;
|
|
+ writel(val, csr_base + BRIDGE_CTRL_1);
|
|
+}
|
|
+
|
|
+static u64 xgene_pcie_set_ib_mask(void __iomem *csr_base, u32 addr,
|
|
+ u32 flags, u64 size)
|
|
+{
|
|
+ u64 mask = (~(size - 1) & PCI_BASE_ADDRESS_MEM_MASK) | flags;
|
|
+ u32 val32 = 0;
|
|
+ u32 val;
|
|
+
|
|
+ val32 = readl(csr_base + addr);
|
|
+ val = (val32 & 0x0000ffff) | (lower_32_bits(mask) << 16);
|
|
+ writel(val, csr_base + addr);
|
|
+
|
|
+ val32 = readl(csr_base + addr + 0x04);
|
|
+ val = (val32 & 0xffff0000) | (lower_32_bits(mask) >> 16);
|
|
+ writel(val, csr_base + addr + 0x04);
|
|
+
|
|
+ val32 = readl(csr_base + addr + 0x04);
|
|
+ val = (val32 & 0x0000ffff) | (upper_32_bits(mask) << 16);
|
|
+ writel(val, csr_base + addr + 0x04);
|
|
+
|
|
+ val32 = readl(csr_base + addr + 0x08);
|
|
+ val = (val32 & 0xffff0000) | (upper_32_bits(mask) >> 16);
|
|
+ writel(val, csr_base + addr + 0x08);
|
|
+
|
|
+ return mask;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_poll_linkup(struct xgene_pcie_port *port,
|
|
+ u32 *lanes, u32 *speed)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ ulong timeout;
|
|
+ u32 val32;
|
|
+
|
|
+ /*
|
|
+ * A component enters the LTSSM Detect state within
|
|
+ * 20ms of the end of fundamental core reset.
|
|
+ */
|
|
+ msleep(XGENE_LTSSM_DETECT_WAIT);
|
|
+ port->link_up = 0;
|
|
+ timeout = jiffies + msecs_to_jiffies(XGENE_LTSSM_L0_WAIT);
|
|
+ while (time_before(jiffies, timeout)) {
|
|
+ val32 = readl(csr_base + PCIECORE_CTLANDSTATUS);
|
|
+ if (val32 & LINK_UP_MASK) {
|
|
+ port->link_up = 1;
|
|
+ *speed = PIPE_PHY_RATE_RD(val32);
|
|
+ val32 = readl(csr_base + BRIDGE_STATUS_0);
|
|
+ *lanes = val32 >> 26;
|
|
+ break;
|
|
+ }
|
|
+ msleep(1);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_root_complex(struct xgene_pcie_port *port)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ u32 val;
|
|
+
|
|
+ val = (XGENE_PCIE_DEVICEID << 16) | XGENE_PCIE_VENDORID;
|
|
+ writel(val, csr_base + BRIDGE_CFG_0);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_1);
|
|
+ val &= ~CLASS_CODE_MASK;
|
|
+ val |= PCI_CLASS_BRIDGE_PCI << 16;
|
|
+ writel(val, csr_base + BRIDGE_CFG_1);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_14);
|
|
+ val |= SWITCH_PORT_MODE_MASK;
|
|
+ val &= ~PM_FORCE_RP_MODE_MASK;
|
|
+ writel(val, csr_base + BRIDGE_CFG_14);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CTRL_5);
|
|
+ val &= ~DEVICE_PORT_TYPE_MASK;
|
|
+ val |= XGENE_PORT_TYPE_RC;
|
|
+ writel(val, csr_base + BRIDGE_CTRL_5);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CTRL_2);
|
|
+ val |= ENABLE_ASPM;
|
|
+ writel(val, csr_base + BRIDGE_CTRL_2);
|
|
+
|
|
+ val = readl(csr_base + BRIDGE_CFG_32);
|
|
+ writel(val | (1 << 19), csr_base + BRIDGE_CFG_32);
|
|
+}
|
|
+
|
|
+/* Return 0 on success */
|
|
+static int xgene_pcie_init_ecc(struct xgene_pcie_port *port)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ ulong timeout;
|
|
+ u32 val;
|
|
+
|
|
+ val = readl(csr_base + MEM_RAM_SHUTDOWN);
|
|
+ if (!val)
|
|
+ return 0;
|
|
+ writel(0x0, csr_base + MEM_RAM_SHUTDOWN);
|
|
+ timeout = jiffies + msecs_to_jiffies(XGENE_PCIE_ECC_TIMEOUT);
|
|
+ while (time_before(jiffies, timeout)) {
|
|
+ val = readl(csr_base + BLOCK_MEM_RDY);
|
|
+ if (val == BLOCK_MEM_RDY_VAL)
|
|
+ return 0;
|
|
+ msleep(1);
|
|
+ }
|
|
+
|
|
+ return 1;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_init_port(struct xgene_pcie_port *port)
|
|
+{
|
|
+ int rc;
|
|
+
|
|
+ port->clk = clk_get(port->dev, NULL);
|
|
+ if (IS_ERR(port->clk)) {
|
|
+ dev_err(port->dev, "clock not available\n");
|
|
+ return -ENODEV;
|
|
+ }
|
|
+
|
|
+ rc = clk_prepare_enable(port->clk);
|
|
+ if (rc) {
|
|
+ dev_err(port->dev, "clock enable failed\n");
|
|
+ return rc;
|
|
+ }
|
|
+
|
|
+ rc = xgene_pcie_init_ecc(port);
|
|
+ if (rc) {
|
|
+ dev_err(port->dev, "memory init failed\n");
|
|
+ return rc;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_fixup_bridge(struct pci_dev *dev)
|
|
+{
|
|
+ int i;
|
|
+
|
|
+ /* Hide the PCI host BARs from the kernel as their content doesn't
|
|
+ * fit well in the resource management
|
|
+ */
|
|
+ for (i = 0; i < DEVICE_COUNT_RESOURCE; i++) {
|
|
+ dev->resource[i].start = dev->resource[i].end = 0;
|
|
+ dev->resource[i].flags = 0;
|
|
+ }
|
|
+ dev_info(&dev->dev, "Hiding X-Gene pci host bridge resources %s\n",
|
|
+ pci_name(dev));
|
|
+}
|
|
+DECLARE_PCI_FIXUP_HEADER(XGENE_PCIE_VENDORID, XGENE_PCIE_DEVICEID,
|
|
+ xgene_pcie_fixup_bridge);
|
|
+
|
|
+static int xgene_pcie_map_reg(struct xgene_pcie_port *port,
|
|
+ struct platform_device *pdev, u64 *cfg_addr)
|
|
+{
|
|
+ struct resource *res;
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "csr");
|
|
+ port->csr_base = devm_ioremap_resource(port->dev, res);
|
|
+ if (IS_ERR(port->csr_base))
|
|
+ return PTR_ERR(port->csr_base);
|
|
+
|
|
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "cfg");
|
|
+ port->cfg_base = devm_ioremap_resource(port->dev, res);
|
|
+ if (IS_ERR(port->cfg_base))
|
|
+ return PTR_ERR(port->cfg_base);
|
|
+ *cfg_addr = res->start;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_ob_reg(struct xgene_pcie_port *port,
|
|
+ struct resource *res, u32 offset, u64 addr)
|
|
+{
|
|
+ void __iomem *base = port->csr_base + offset;
|
|
+ resource_size_t size = resource_size(res);
|
|
+ u64 restype = resource_type(res);
|
|
+ u64 cpu_addr, pci_addr;
|
|
+ u64 mask = 0;
|
|
+ u32 min_size;
|
|
+ u32 flag = EN_REG;
|
|
+
|
|
+ if (restype == IORESOURCE_MEM) {
|
|
+ cpu_addr = res->start;
|
|
+ pci_addr = addr;
|
|
+ min_size = SZ_128M;
|
|
+ } else {
|
|
+ cpu_addr = addr;
|
|
+ pci_addr = res->start;
|
|
+ min_size = 128;
|
|
+ flag |= OB_LO_IO;
|
|
+ }
|
|
+ if (size >= min_size)
|
|
+ mask = ~(size - 1) | flag;
|
|
+ else
|
|
+ dev_warn(port->dev, "res size 0x%llx less than minimum 0x%x\n",
|
|
+ (u64)size, min_size);
|
|
+ writel(lower_32_bits(cpu_addr), base);
|
|
+ writel(upper_32_bits(cpu_addr), base + 0x04);
|
|
+ writel(lower_32_bits(mask), base + 0x08);
|
|
+ writel(upper_32_bits(mask), base + 0x0c);
|
|
+ writel(lower_32_bits(pci_addr), base + 0x10);
|
|
+ writel(upper_32_bits(pci_addr), base + 0x14);
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_cfg_reg(void __iomem *csr_base, u64 addr)
|
|
+{
|
|
+ writel(lower_32_bits(addr), csr_base + CFGBARL);
|
|
+ writel(upper_32_bits(addr), csr_base + CFGBARH);
|
|
+ writel(EN_REG, csr_base + CFGCTL);
|
|
+}
|
|
+
|
|
+static int xgene_pcie_map_ranges(struct xgene_pcie_port *port,
|
|
+ struct pci_host_bridge *bridge,
|
|
+ u64 cfg_addr)
|
|
+{
|
|
+ struct device *dev = port->dev;
|
|
+ struct pci_host_bridge_window *window;
|
|
+ int ret;
|
|
+
|
|
+ list_for_each_entry(window, &bridge->windows, list) {
|
|
+ struct resource *res = window->res;
|
|
+ u64 restype = resource_type(res);
|
|
+ dev_dbg(port->dev, "0x%08lx 0x%016llx...0x%016llx\n",
|
|
+ res->flags, res->start, res->end);
|
|
+
|
|
+ switch (restype) {
|
|
+ case IORESOURCE_IO:
|
|
+ xgene_pcie_setup_ob_reg(port, res, OMR2BARL,
|
|
+ bridge->io_base);
|
|
+ ret = pci_remap_iospace(res, bridge->io_base);
|
|
+ if (ret < 0)
|
|
+ return ret;
|
|
+ break;
|
|
+ case IORESOURCE_MEM:
|
|
+ xgene_pcie_setup_ob_reg(port, res, OMR3BARL,
|
|
+ res->start - window->offset);
|
|
+ break;
|
|
+ case IORESOURCE_BUS:
|
|
+ break;
|
|
+ default:
|
|
+ dev_err(dev, "invalid io resource!");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+ }
|
|
+ xgene_pcie_setup_cfg_reg(port->csr_base, cfg_addr);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_pims(void *addr, u64 pim, u64 size)
|
|
+{
|
|
+ writel(lower_32_bits(pim), addr);
|
|
+ writel(upper_32_bits(pim) | EN_COHERENCY, addr + 0x04);
|
|
+ writel(lower_32_bits(size), addr + 0x10);
|
|
+ writel(upper_32_bits(size), addr + 0x14);
|
|
+}
|
|
+
|
|
+/*
|
|
+ * X-Gene PCIe support maximum 3 inbound memory regions
|
|
+ * This function helps to select a region based on size of region
|
|
+ */
|
|
+static int xgene_pcie_select_ib_reg(u8 *ib_reg_mask, u64 size)
|
|
+{
|
|
+ if ((size > 4) && (size < SZ_16M) && !(*ib_reg_mask & (1 << 1))) {
|
|
+ *ib_reg_mask |= (1 << 1);
|
|
+ return 1;
|
|
+ }
|
|
+
|
|
+ if ((size > SZ_1K) && (size < SZ_1T) && !(*ib_reg_mask & (1 << 0))) {
|
|
+ *ib_reg_mask |= (1 << 0);
|
|
+ return 0;
|
|
+ }
|
|
+
|
|
+ if ((size > SZ_1M) && (size < SZ_1T) && !(*ib_reg_mask & (1 << 2))) {
|
|
+ *ib_reg_mask |= (1 << 2);
|
|
+ return 2;
|
|
+ }
|
|
+ return -EINVAL;
|
|
+}
|
|
+
|
|
+static void xgene_pcie_setup_ib_reg(struct xgene_pcie_port *port,
|
|
+ struct of_pci_range *range, u8 *ib_reg_mask)
|
|
+{
|
|
+ void __iomem *csr_base = port->csr_base;
|
|
+ void __iomem *cfg_base = port->cfg_base;
|
|
+ void *bar_addr;
|
|
+ void *pim_addr;
|
|
+ u64 restype = range->flags & IORESOURCE_TYPE_BITS;
|
|
+ u64 cpu_addr = range->cpu_addr;
|
|
+ u64 pci_addr = range->pci_addr;
|
|
+ u64 size = range->size;
|
|
+ u64 mask = ~(size - 1) | EN_REG;
|
|
+ u32 flags = PCI_BASE_ADDRESS_MEM_TYPE_64;
|
|
+ u32 bar_low;
|
|
+ int region;
|
|
+
|
|
+ region = xgene_pcie_select_ib_reg(ib_reg_mask, range->size);
|
|
+ if (region < 0) {
|
|
+ dev_warn(port->dev, "invalid pcie dma-range config\n");
|
|
+ return;
|
|
+ }
|
|
+
|
|
+ if (restype == PCI_BASE_ADDRESS_MEM_PREFETCH)
|
|
+ flags |= PCI_BASE_ADDRESS_MEM_PREFETCH;
|
|
+
|
|
+ bar_low = pcie_bar_low_val((u32)cpu_addr, flags);
|
|
+ switch (region) {
|
|
+ case 0:
|
|
+ xgene_pcie_set_ib_mask(csr_base, BRIDGE_CFG_4, flags, size);
|
|
+ bar_addr = cfg_base + PCI_BASE_ADDRESS_0;
|
|
+ writel(bar_low, bar_addr);
|
|
+ writel(upper_32_bits(cpu_addr), bar_addr + 0x4);
|
|
+ pim_addr = csr_base + PIM1_1L;
|
|
+ break;
|
|
+ case 1:
|
|
+ bar_addr = csr_base + IBAR2;
|
|
+ writel(bar_low, bar_addr);
|
|
+ writel(lower_32_bits(mask), csr_base + IR2MSK);
|
|
+ pim_addr = csr_base + PIM2_1L;
|
|
+ break;
|
|
+ case 2:
|
|
+ bar_addr = csr_base + IBAR3L;
|
|
+ writel(bar_low, bar_addr);
|
|
+ writel(upper_32_bits(cpu_addr), bar_addr + 0x4);
|
|
+ writel(lower_32_bits(mask), csr_base + IR3MSKL);
|
|
+ writel(upper_32_bits(mask), csr_base + IR3MSKL + 0x4);
|
|
+ pim_addr = csr_base + PIM3_1L;
|
|
+ break;
|
|
+ }
|
|
+
|
|
+ xgene_pcie_setup_pims(pim_addr, pci_addr, size);
|
|
+}
|
|
+
|
|
+static int pci_dma_range_parser_init(struct of_pci_range_parser *parser,
|
|
+ struct device_node *node)
|
|
+{
|
|
+ const int na = 3, ns = 2;
|
|
+ int rlen;
|
|
+
|
|
+ parser->node = node;
|
|
+ parser->pna = of_n_addr_cells(node);
|
|
+ parser->np = parser->pna + na + ns;
|
|
+
|
|
+ parser->range = of_get_property(node, "dma-ranges", &rlen);
|
|
+ if (!parser->range)
|
|
+ return -ENOENT;
|
|
+
|
|
+ parser->end = parser->range + rlen / sizeof(__be32);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_parse_map_dma_ranges(struct xgene_pcie_port *port)
|
|
+{
|
|
+ struct device_node *np = port->node;
|
|
+ struct of_pci_range range;
|
|
+ struct of_pci_range_parser parser;
|
|
+ struct device *dev = port->dev;
|
|
+ u8 ib_reg_mask = 0;
|
|
+
|
|
+ if (pci_dma_range_parser_init(&parser, np)) {
|
|
+ dev_err(dev, "missing dma-ranges property\n");
|
|
+ return -EINVAL;
|
|
+ }
|
|
+
|
|
+ /* Get the dma-ranges from DT */
|
|
+ for_each_of_pci_range(&parser, &range) {
|
|
+ u64 end = range.cpu_addr + range.size - 1;
|
|
+ dev_dbg(port->dev, "0x%08x 0x%016llx..0x%016llx -> 0x%016llx\n",
|
|
+ range.flags, range.cpu_addr, end, range.pci_addr);
|
|
+ xgene_pcie_setup_ib_reg(port, &range, &ib_reg_mask);
|
|
+ }
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static int xgene_pcie_probe_bridge(struct platform_device *pdev)
|
|
+{
|
|
+ struct device_node *np = of_node_get(pdev->dev.of_node);
|
|
+ struct xgene_pcie_port *port;
|
|
+ struct pci_host_bridge *bridge;
|
|
+ resource_size_t lastbus;
|
|
+ u32 lanes = 0, speed = 0;
|
|
+ u64 cfg_addr = 0;
|
|
+ int ret;
|
|
+
|
|
+ port = devm_kzalloc(&pdev->dev, sizeof(*port), GFP_KERNEL);
|
|
+ if (!port)
|
|
+ return -ENOMEM;
|
|
+ port->node = np;
|
|
+ port->dev = &pdev->dev;
|
|
+
|
|
+ ret = xgene_pcie_map_reg(port, pdev, &cfg_addr);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ ret = xgene_pcie_init_port(port);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+ xgene_pcie_program_core(port->csr_base);
|
|
+ xgene_pcie_setup_root_complex(port);
|
|
+
|
|
+ bridge = of_create_pci_host_bridge(&pdev->dev, &xgene_pcie_ops, port);
|
|
+ if (IS_ERR_OR_NULL(bridge))
|
|
+ return PTR_ERR(bridge);
|
|
+
|
|
+ ret = xgene_pcie_map_ranges(port, bridge, cfg_addr);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ ret = xgene_pcie_parse_map_dma_ranges(port);
|
|
+ if (ret)
|
|
+ return ret;
|
|
+
|
|
+ xgene_pcie_poll_linkup(port, &lanes, &speed);
|
|
+ if (!port->link_up)
|
|
+ dev_info(port->dev, "(rc) link down\n");
|
|
+ else
|
|
+ dev_info(port->dev, "(rc) x%d gen-%d link up\n",
|
|
+ lanes, speed + 1);
|
|
+ platform_set_drvdata(pdev, port);
|
|
+ lastbus = pci_rescan_bus(bridge->bus);
|
|
+ pci_bus_update_busn_res_end(bridge->bus, lastbus);
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static const struct of_device_id xgene_pcie_match_table[] = {
|
|
+ {.compatible = "apm,xgene-pcie",},
|
|
+ {},
|
|
+};
|
|
+
|
|
+static struct platform_driver xgene_pcie_driver = {
|
|
+ .driver = {
|
|
+ .name = "xgene-pcie",
|
|
+ .owner = THIS_MODULE,
|
|
+ .of_match_table = of_match_ptr(xgene_pcie_match_table),
|
|
+ },
|
|
+ .probe = xgene_pcie_probe_bridge,
|
|
+};
|
|
+module_platform_driver(xgene_pcie_driver);
|
|
+
|
|
+MODULE_AUTHOR("Tanmay Inamdar <tinamdar@apm.com>");
|
|
+MODULE_DESCRIPTION("APM X-Gene PCIe driver");
|
|
+MODULE_LICENSE("GPL v2");
|
|
diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c
|
|
index 2c9ac70..7bae0f9 100644
|
|
--- a/drivers/pci/pci.c
|
|
+++ b/drivers/pci/pci.c
|
|
@@ -17,6 +17,7 @@
|
|
#include <linux/spinlock.h>
|
|
#include <linux/string.h>
|
|
#include <linux/log2.h>
|
|
+#include <linux/of_pci.h>
|
|
#include <linux/pci-aspm.h>
|
|
#include <linux/pm_wakeup.h>
|
|
#include <linux/interrupt.h>
|
|
@@ -1453,6 +1454,9 @@ EXPORT_SYMBOL(pcim_pin_device);
|
|
*/
|
|
int __weak pcibios_add_device(struct pci_dev *dev)
|
|
{
|
|
+#ifdef CONFIG_OF
|
|
+ dev->irq = of_irq_parse_and_map_pci(dev, 0, 0);
|
|
+#endif
|
|
return 0;
|
|
}
|
|
|
|
@@ -2704,6 +2708,39 @@ int pci_request_regions_exclusive(struct pci_dev *pdev, const char *res_name)
|
|
}
|
|
EXPORT_SYMBOL(pci_request_regions_exclusive);
|
|
|
|
+/**
|
|
+ * pci_remap_iospace - Remap the memory mapped I/O space
|
|
+ * @res: Resource describing the I/O space
|
|
+ * @phys_addr: physical address where the range will be mapped.
|
|
+ *
|
|
+ * Remap the memory mapped I/O space described by the @res
|
|
+ * into the CPU physical address space. Only architectures
|
|
+ * that have memory mapped IO defined (and hence PCI_IOBASE)
|
|
+ * should call this function.
|
|
+ */
|
|
+int __weak pci_remap_iospace(const struct resource *res, phys_addr_t phys_addr)
|
|
+{
|
|
+ int err = -ENODEV;
|
|
+
|
|
+#ifdef PCI_IOBASE
|
|
+ if (!(res->flags & IORESOURCE_IO))
|
|
+ return -EINVAL;
|
|
+
|
|
+ if (res->end > IO_SPACE_LIMIT)
|
|
+ return -EINVAL;
|
|
+
|
|
+ err = ioremap_page_range(res->start + (unsigned long)PCI_IOBASE,
|
|
+ res->end + 1 + (unsigned long)PCI_IOBASE,
|
|
+ phys_addr, __pgprot(PROT_DEVICE_nGnRE));
|
|
+#else
|
|
+ /* this architecture does not have memory mapped I/O space,
|
|
+ so this function should never be called */
|
|
+ WARN_ON(1);
|
|
+#endif
|
|
+
|
|
+ return err;
|
|
+}
|
|
+
|
|
static void __pci_set_master(struct pci_dev *dev, bool enable)
|
|
{
|
|
u16 old_cmd, cmd;
|
|
diff --git a/drivers/pci/probe.c b/drivers/pci/probe.c
|
|
index e3cf8a2..abf5e82 100644
|
|
--- a/drivers/pci/probe.c
|
|
+++ b/drivers/pci/probe.c
|
|
@@ -515,7 +515,7 @@ static void pci_release_host_bridge_dev(struct device *dev)
|
|
kfree(bridge);
|
|
}
|
|
|
|
-static struct pci_host_bridge *pci_alloc_host_bridge(struct pci_bus *b)
|
|
+static struct pci_host_bridge *pci_alloc_host_bridge(void)
|
|
{
|
|
struct pci_host_bridge *bridge;
|
|
|
|
@@ -524,7 +524,6 @@ static struct pci_host_bridge *pci_alloc_host_bridge(struct pci_bus *b)
|
|
return NULL;
|
|
|
|
INIT_LIST_HEAD(&bridge->windows);
|
|
- bridge->bus = b;
|
|
return bridge;
|
|
}
|
|
|
|
@@ -1749,8 +1748,9 @@ void __weak pcibios_remove_bus(struct pci_bus *bus)
|
|
{
|
|
}
|
|
|
|
-struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
- struct pci_ops *ops, void *sysdata, struct list_head *resources)
|
|
+struct pci_bus *pci_create_root_bus_in_domain(struct device *parent,
|
|
+ int domain, int bus, struct pci_ops *ops, void *sysdata,
|
|
+ struct list_head *resources)
|
|
{
|
|
int error;
|
|
struct pci_host_bridge *bridge;
|
|
@@ -1761,37 +1761,41 @@ struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
char bus_addr[64];
|
|
char *fmt;
|
|
|
|
+ bridge = pci_alloc_host_bridge();
|
|
+ if (!bridge)
|
|
+ return ERR_PTR(-ENOMEM);
|
|
+
|
|
+ bridge->dev.parent = parent;
|
|
+ bridge->dev.release = pci_release_host_bridge_dev;
|
|
+ bridge->domain_nr = domain;
|
|
+
|
|
b = pci_alloc_bus();
|
|
- if (!b)
|
|
- return NULL;
|
|
+ if (!b) {
|
|
+ error = -ENOMEM;
|
|
+ goto err_out;
|
|
+ }
|
|
|
|
b->sysdata = sysdata;
|
|
b->ops = ops;
|
|
b->number = b->busn_res.start = bus;
|
|
- b2 = pci_find_bus(pci_domain_nr(b), bus);
|
|
+ b2 = pci_find_bus(bridge->domain_nr, bus);
|
|
if (b2) {
|
|
/* If we already got to this bus through a different bridge, ignore it */
|
|
dev_dbg(&b2->dev, "bus already known\n");
|
|
- goto err_out;
|
|
+ error = -EEXIST;
|
|
+ goto err_bus_out;
|
|
}
|
|
|
|
- bridge = pci_alloc_host_bridge(b);
|
|
- if (!bridge)
|
|
- goto err_out;
|
|
-
|
|
- bridge->dev.parent = parent;
|
|
- bridge->dev.release = pci_release_host_bridge_dev;
|
|
- dev_set_name(&bridge->dev, "pci%04x:%02x", pci_domain_nr(b), bus);
|
|
+ bridge->bus = b;
|
|
+ dev_set_name(&bridge->dev, "pci%04x:%02x", bridge->domain_nr, bus);
|
|
error = pcibios_root_bridge_prepare(bridge);
|
|
- if (error) {
|
|
- kfree(bridge);
|
|
+ if (error)
|
|
goto err_out;
|
|
- }
|
|
|
|
error = device_register(&bridge->dev);
|
|
if (error) {
|
|
put_device(&bridge->dev);
|
|
- goto err_out;
|
|
+ goto err_bus_out;
|
|
}
|
|
b->bridge = get_device(&bridge->dev);
|
|
device_enable_async_suspend(b->bridge);
|
|
@@ -1802,7 +1806,7 @@ struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
|
|
b->dev.class = &pcibus_class;
|
|
b->dev.parent = b->bridge;
|
|
- dev_set_name(&b->dev, "%04x:%02x", pci_domain_nr(b), bus);
|
|
+ dev_set_name(&b->dev, "%04x:%02x", bridge->domain_nr, bus);
|
|
error = device_register(&b->dev);
|
|
if (error)
|
|
goto class_dev_reg_err;
|
|
@@ -1848,9 +1852,31 @@ struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
class_dev_reg_err:
|
|
put_device(&bridge->dev);
|
|
device_unregister(&bridge->dev);
|
|
+err_bus_out:
|
|
+ kfree(b);
|
|
err_out:
|
|
+ kfree(bridge);
|
|
+ return ERR_PTR(error);
|
|
+}
|
|
+
|
|
+struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
+ struct pci_ops *ops, void *sysdata, struct list_head *resources)
|
|
+{
|
|
+ int domain_nr;
|
|
+ struct pci_bus *b = pci_alloc_bus();
|
|
+ if (!b)
|
|
+ return NULL;
|
|
+
|
|
+ b->sysdata = sysdata;
|
|
+ domain_nr = pci_domain_nr(b);
|
|
kfree(b);
|
|
- return NULL;
|
|
+
|
|
+ b = pci_create_root_bus_in_domain(parent, domain_nr, bus,
|
|
+ ops, sysdata, resources);
|
|
+ if (IS_ERR(b))
|
|
+ return NULL;
|
|
+
|
|
+ return b;
|
|
}
|
|
|
|
int pci_bus_insert_busn_res(struct pci_bus *b, int bus, int bus_max)
|
|
diff --git a/include/asm-generic/io.h b/include/asm-generic/io.h
|
|
index 975e1cc..2e2161b 100644
|
|
--- a/include/asm-generic/io.h
|
|
+++ b/include/asm-generic/io.h
|
|
@@ -331,7 +331,7 @@ static inline void iounmap(void __iomem *addr)
|
|
#ifndef CONFIG_GENERIC_IOMAP
|
|
static inline void __iomem *ioport_map(unsigned long port, unsigned int nr)
|
|
{
|
|
- return (void __iomem *) port;
|
|
+ return (void __iomem *)(PCI_IOBASE + (port & IO_SPACE_LIMIT));
|
|
}
|
|
|
|
static inline void ioport_unmap(void __iomem *p)
|
|
diff --git a/include/linux/of_address.h b/include/linux/of_address.h
|
|
index fb7b722..cb9479e4 100644
|
|
--- a/include/linux/of_address.h
|
|
+++ b/include/linux/of_address.h
|
|
@@ -23,17 +23,8 @@ struct of_pci_range {
|
|
#define for_each_of_pci_range(parser, range) \
|
|
for (; of_pci_range_parser_one(parser, range);)
|
|
|
|
-static inline void of_pci_range_to_resource(struct of_pci_range *range,
|
|
- struct device_node *np,
|
|
- struct resource *res)
|
|
-{
|
|
- res->flags = range->flags;
|
|
- res->start = range->cpu_addr;
|
|
- res->end = range->cpu_addr + range->size - 1;
|
|
- res->parent = res->child = res->sibling = NULL;
|
|
- res->name = np->full_name;
|
|
-}
|
|
-
|
|
+extern int of_pci_range_to_resource(struct of_pci_range *range,
|
|
+ struct device_node *np, struct resource *res);
|
|
/* Translate a DMA address from device space to CPU space */
|
|
extern u64 of_translate_dma_address(struct device_node *dev,
|
|
const __be32 *in_addr);
|
|
@@ -55,6 +46,7 @@ extern void __iomem *of_iomap(struct device_node *device, int index);
|
|
extern const __be32 *of_get_address(struct device_node *dev, int index,
|
|
u64 *size, unsigned int *flags);
|
|
|
|
+extern int pci_register_io_range(phys_addr_t addr, resource_size_t size);
|
|
extern unsigned long pci_address_to_pio(phys_addr_t addr);
|
|
|
|
extern int of_pci_range_parser_init(struct of_pci_range_parser *parser,
|
|
diff --git a/include/linux/of_pci.h b/include/linux/of_pci.h
|
|
index dde3a4a..71e36d0 100644
|
|
--- a/include/linux/of_pci.h
|
|
+++ b/include/linux/of_pci.h
|
|
@@ -15,6 +15,9 @@ struct device_node *of_pci_find_child_device(struct device_node *parent,
|
|
int of_pci_get_devfn(struct device_node *np);
|
|
int of_irq_parse_and_map_pci(const struct pci_dev *dev, u8 slot, u8 pin);
|
|
int of_pci_parse_bus_range(struct device_node *node, struct resource *res);
|
|
+struct pci_host_bridge *of_create_pci_host_bridge(struct device *parent,
|
|
+ struct pci_ops *ops, void *host_data);
|
|
+
|
|
#else
|
|
static inline int of_irq_parse_pci(const struct pci_dev *pdev, struct of_phandle_args *out_irq)
|
|
{
|
|
@@ -43,6 +46,13 @@ of_pci_parse_bus_range(struct device_node *node, struct resource *res)
|
|
{
|
|
return -EINVAL;
|
|
}
|
|
+
|
|
+static inline struct pci_host_bridge *
|
|
+of_create_pci_host_bridge(struct device *parent, struct pci_ops *ops,
|
|
+ void *host_data)
|
|
+{
|
|
+ return NULL;
|
|
+}
|
|
#endif
|
|
|
|
#if defined(CONFIG_OF) && defined(CONFIG_PCI_MSI)
|
|
diff --git a/include/linux/pci.h b/include/linux/pci.h
|
|
index 61978a4..f582746 100644
|
|
--- a/include/linux/pci.h
|
|
+++ b/include/linux/pci.h
|
|
@@ -401,6 +401,8 @@ struct pci_host_bridge_window {
|
|
struct pci_host_bridge {
|
|
struct device dev;
|
|
struct pci_bus *bus; /* root bus */
|
|
+ int domain_nr;
|
|
+ resource_size_t io_base; /* physical address for the start of I/O area */
|
|
struct list_head windows; /* pci_host_bridge_windows */
|
|
void (*release_fn)(struct pci_host_bridge *);
|
|
void *release_data;
|
|
@@ -769,6 +771,9 @@ struct pci_bus *pci_scan_bus(int bus, struct pci_ops *ops, void *sysdata);
|
|
struct pci_bus *pci_create_root_bus(struct device *parent, int bus,
|
|
struct pci_ops *ops, void *sysdata,
|
|
struct list_head *resources);
|
|
+struct pci_bus *pci_create_root_bus_in_domain(struct device *parent,
|
|
+ int domain, int bus, struct pci_ops *ops,
|
|
+ void *sysdata, struct list_head *resources);
|
|
int pci_bus_insert_busn_res(struct pci_bus *b, int bus, int busmax);
|
|
int pci_bus_update_busn_res_end(struct pci_bus *b, int busmax);
|
|
void pci_bus_release_busn_res(struct pci_bus *b);
|
|
@@ -1097,6 +1102,9 @@ int __must_check pci_bus_alloc_resource(struct pci_bus *bus,
|
|
resource_size_t),
|
|
void *alignf_data);
|
|
|
|
+
|
|
+int pci_remap_iospace(const struct resource *res, phys_addr_t phys_addr);
|
|
+
|
|
static inline dma_addr_t pci_bus_address(struct pci_dev *pdev, int bar)
|
|
{
|
|
struct pci_bus_region region;
|
|
@@ -1815,8 +1823,15 @@ static inline void pci_set_of_node(struct pci_dev *dev) { }
|
|
static inline void pci_release_of_node(struct pci_dev *dev) { }
|
|
static inline void pci_set_bus_of_node(struct pci_bus *bus) { }
|
|
static inline void pci_release_bus_of_node(struct pci_bus *bus) { }
|
|
+
|
|
#endif /* CONFIG_OF */
|
|
|
|
+/* Used by architecture code to apply any quirks to the list of
|
|
+ * pci_host_bridge resource ranges before they are being used
|
|
+ * by of_create_pci_host_bridge()
|
|
+ */
|
|
+extern int pcibios_fixup_bridge_ranges(struct list_head *resources);
|
|
+
|
|
#ifdef CONFIG_EEH
|
|
static inline struct eeh_dev *pci_dev_to_eeh_dev(struct pci_dev *pdev)
|
|
{
|
|
diff --git a/tools/perf/arch/arm64/include/perf_regs.h b/tools/perf/arch/arm64/include/perf_regs.h
|
|
index e9441b9..1d3f39c 100644
|
|
--- a/tools/perf/arch/arm64/include/perf_regs.h
|
|
+++ b/tools/perf/arch/arm64/include/perf_regs.h
|
|
@@ -6,6 +6,8 @@
|
|
#include <asm/perf_regs.h>
|
|
|
|
#define PERF_REGS_MASK ((1ULL << PERF_REG_ARM64_MAX) - 1)
|
|
+#define PERF_REGS_MAX PERF_REG_ARM64_MAX
|
|
+
|
|
#define PERF_REG_IP PERF_REG_ARM64_PC
|
|
#define PERF_REG_SP PERF_REG_ARM64_SP
|
|
|