blob: 1f12ae54e9a9946344c7a25e3a621977725155a1 [file] [log] [blame]
// Copyright 2018 The Fuchsia Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "s905d2-gpu.h"
#include "s912-gpu.h"
#include "t931-gpu.h"
#include <ddk/binding.h>
#include <ddk/debug.h>
#include <ddk/device.h>
#include <ddk/driver.h>
#include <ddk/platform-defs.h>
#include <ddk/protocol/iommu.h>
#include <ddk/protocol/platform-device-lib.h>
#include <ddk/protocol/platform/bus.h>
#include <ddk/protocol/platform/device.h>
#include <fuchsia/hardware/gpu/clock/c/fidl.h>
#include <hw/reg.h>
#include <stdint.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <unistd.h>
#include <zircon/process.h>
#include <zircon/syscalls.h>
zx_status_t aml_gp0_init(aml_gpu_t* gpu);
void aml_gp0_release(aml_gpu_t* gpu);
static int32_t current_clk_source;
static void aml_gpu_set_clk_freq_source(aml_gpu_t* gpu, int32_t clk_source) {
if (current_clk_source == clk_source) {
return;
}
aml_gpu_block_t* gpu_block = gpu->gpu_block;
GPU_INFO("Setting clock source to %d: %d\n", clk_source, gpu_block->gpu_clk_freq[clk_source]);
uint32_t current_clk_cntl = READ32_HIU_REG(gpu_block->hhi_clock_cntl_offset);
uint32_t enabled_mux = current_clk_cntl & (1 << FINAL_MUX_BIT_SHIFT);
uint32_t new_mux = enabled_mux == 0;
uint32_t mux_shift = new_mux ? 16 : 0;
// clear existing values
current_clk_cntl &= ~(CLOCK_MUX_MASK << mux_shift);
// set the divisor, enable & source for the unused mux
current_clk_cntl |= CALCULATE_CLOCK_MUX(true,
gpu_block->gpu_clk_freq[clk_source], 1)
<< mux_shift;
// Write the new values to the unused mux
WRITE32_HIU_REG(gpu_block->hhi_clock_cntl_offset, current_clk_cntl);
zx_nanosleep(zx_deadline_after(ZX_USEC(10)));
// Toggle current mux selection
current_clk_cntl ^= (1 << FINAL_MUX_BIT_SHIFT);
// Select the unused input mux
WRITE32_HIU_REG(gpu_block->hhi_clock_cntl_offset, current_clk_cntl);
current_clk_source = clk_source;
}
static void aml_gpu_set_initial_clk_freq_source(aml_gpu_t* gpu, int32_t clk_source) {
aml_gpu_block_t* gpu_block = gpu->gpu_block;
uint32_t current_clk_cntl = READ32_HIU_REG(gpu_block->hhi_clock_cntl_offset);
uint32_t enabled_mux = (current_clk_cntl & (1 << FINAL_MUX_BIT_SHIFT)) != 0;
uint32_t mux_shift = enabled_mux ? 16 : 0;
if (current_clk_cntl & (1 << (mux_shift + CLK_ENABLED_BIT_SHIFT))) {
aml_gpu_set_clk_freq_source(gpu, clk_source);
} else {
GPU_INFO("Setting initial clock source to %d: %d\n", clk_source, gpu_block->gpu_clk_freq[clk_source]);
// Switching the final dynamic mux from a disabled source to an enabled
// source doesn't work. If the current clock source is disabled, then
// enable it instead of switching.
current_clk_cntl &= ~(CLOCK_MUX_MASK << mux_shift);
current_clk_cntl |= CALCULATE_CLOCK_MUX(true,
gpu_block->gpu_clk_freq[clk_source], 1)
<< mux_shift;
// Write the new values to the existing mux.
WRITE32_HIU_REG(gpu_block->hhi_clock_cntl_offset, current_clk_cntl);
zx_nanosleep(zx_deadline_after(ZX_USEC(10)));
current_clk_source = clk_source;
}
}
static void aml_gpu_init(aml_gpu_t* gpu) {
uint32_t temp;
aml_gpu_block_t* gpu_block = gpu->gpu_block;
temp = READ32_PRESET_REG(gpu_block->reset0_mask_offset);
temp &= ~(1 << 20);
WRITE32_PRESET_REG(gpu_block->reset0_mask_offset, temp);
temp = READ32_PRESET_REG(gpu_block->reset0_level_offset);
temp &= ~(1 << 20);
WRITE32_PRESET_REG(gpu_block->reset0_level_offset, temp);
temp = READ32_PRESET_REG(gpu_block->reset2_mask_offset);
temp &= ~(1 << 14);
WRITE32_PRESET_REG(gpu_block->reset2_mask_offset, temp);
temp = READ32_PRESET_REG(gpu_block->reset2_level_offset);
temp &= ~(1 << 14);
WRITE32_PRESET_REG(gpu_block->reset2_level_offset, temp);
// Currently the index 2 corresponds to the default
// value of GPU clock freq which is 500Mhz.
// In future, the GPU driver in garnet
// can make an IOCTL to set the default freq
aml_gpu_set_initial_clk_freq_source(gpu, 2);
temp = READ32_PRESET_REG(gpu_block->reset0_level_offset);
temp |= 1 << 20;
WRITE32_PRESET_REG(gpu_block->reset0_level_offset, temp);
temp = READ32_PRESET_REG(gpu_block->reset2_level_offset);
temp |= 1 << 14;
WRITE32_PRESET_REG(gpu_block->reset2_level_offset, temp);
WRITE32_GPU_REG(PWR_KEY, 0x2968A819);
WRITE32_GPU_REG(PWR_OVERRIDE1, 0xfff | (0x20 << 16));
}
static void aml_gpu_release(void* ctx) {
aml_gpu_t* gpu = ctx;
aml_gp0_release(gpu);
mmio_buffer_release(&gpu->hiu_buffer);
mmio_buffer_release(&gpu->preset_buffer);
mmio_buffer_release(&gpu->gpu_buffer);
zx_handle_close(gpu->bti);
free(gpu);
}
static zx_status_t aml_gpu_get_protocol(void* ctx, uint32_t proto_id, void* out_proto) {
aml_gpu_t* gpu = ctx;
pdev_protocol_t* gpu_proto = out_proto;
// Forward the underlying ops.
gpu_proto->ops = gpu->pdev.ops;
gpu_proto->ctx = gpu->pdev.ctx;
return ZX_OK;
}
static zx_status_t aml_gpu_SetFrequencySource(void* ctx, uint32_t clk_source, fidl_txn_t* txn) {
aml_gpu_t* gpu = ctx;
if (clk_source >= MAX_GPU_CLK_FREQ) {
GPU_ERROR("Invalid clock freq source index\n");
return fuchsia_hardware_gpu_clock_ClockSetFrequencySource_reply(txn, ZX_ERR_NOT_SUPPORTED);
}
aml_gpu_set_clk_freq_source(gpu, clk_source);
return fuchsia_hardware_gpu_clock_ClockSetFrequencySource_reply(txn, ZX_OK);
}
static fuchsia_hardware_gpu_clock_Clock_ops_t fidl_ops = {.SetFrequencySource =
aml_gpu_SetFrequencySource};
static zx_status_t aml_gpu_message(void* ctx, fidl_msg_t* msg, fidl_txn_t* txn) {
return fuchsia_hardware_gpu_clock_Clock_dispatch(ctx, txn, msg, &fidl_ops);
}
static zx_protocol_device_t aml_gpu_protocol = {
.version = DEVICE_OPS_VERSION,
.release = aml_gpu_release,
.get_protocol = aml_gpu_get_protocol,
.message = aml_gpu_message,
};
static zx_status_t aml_gpu_bind(void* ctx, zx_device_t* parent) {
zx_status_t status;
aml_gpu_t* gpu = calloc(1, sizeof(aml_gpu_t));
if (!gpu) {
return ZX_ERR_NO_MEMORY;
}
if ((status = device_get_protocol(parent, ZX_PROTOCOL_PDEV, &gpu->pdev)) != ZX_OK) {
GPU_ERROR("ZX_PROTOCOL_PDEV not available\n");
goto fail;
}
status = pdev_get_bti(&gpu->pdev, 0, &gpu->bti);
if (status != ZX_OK) {
GPU_ERROR("could not get BTI handle: %d\n", status);
return status;
}
status = pdev_map_mmio_buffer(&gpu->pdev, MMIO_GPU, ZX_CACHE_POLICY_UNCACHED_DEVICE,
&gpu->gpu_buffer);
if (status != ZX_OK) {
GPU_ERROR("pdev_map_mmio_buffer failed\n");
goto fail;
}
status = pdev_map_mmio_buffer(&gpu->pdev, MMIO_HIU, ZX_CACHE_POLICY_UNCACHED_DEVICE,
&gpu->hiu_buffer);
if (status != ZX_OK) {
GPU_ERROR("pdev_map_mmio_buffer failed\n");
goto fail;
}
status = pdev_map_mmio_buffer(&gpu->pdev, MMIO_PRESET, ZX_CACHE_POLICY_UNCACHED_DEVICE,
&gpu->preset_buffer);
if (status != ZX_OK) {
GPU_ERROR("pdev_map_mmio_buffer failed\n");
goto fail;
}
pdev_device_info_t info;
status = pdev_get_device_info(&gpu->pdev, &info);
if (status != ZX_OK) {
GPU_ERROR("pdev_get_device_info failed\n");
goto fail;
}
switch (info.pid) {
case PDEV_PID_AMLOGIC_S912:
gpu->gpu_block = &s912_gpu_blocks;
break;
case PDEV_PID_AMLOGIC_S905D2:
gpu->gpu_block = &s905d2_gpu_blocks;
break;
case PDEV_PID_AMLOGIC_T931:
gpu->gpu_block = &t931_gpu_blocks;
break;
default:
GPU_ERROR("unsupported SOC PID %u\n", info.pid);
goto fail;
}
if (info.pid == PDEV_PID_AMLOGIC_S905D2) {
status = aml_gp0_init(gpu);
if (status != ZX_OK) {
GPU_ERROR("aml_gp0_init failed: %d\n", status);
goto fail;
}
}
aml_gpu_init(gpu);
zx_device_prop_t props[] = {
{BIND_PROTOCOL, 0, ZX_PROTOCOL_PDEV},
{BIND_PLATFORM_DEV_VID, 0, PDEV_VID_GENERIC},
{BIND_PLATFORM_DEV_PID, 0, PDEV_PID_GENERIC},
{BIND_PLATFORM_DEV_DID, 0, PDEV_DID_ARM_MALI},
};
device_add_args_t args = {
.version = DEVICE_ADD_ARGS_VERSION,
.name = "aml-gpu",
.ctx = gpu,
.ops = &aml_gpu_protocol,
.props = props,
.prop_count = countof(props),
.proto_id = ZX_PROTOCOL_GPU_THERMAL,
};
status = device_add(parent, &args, &gpu->zxdev);
if (status != ZX_OK) {
goto fail;
}
return ZX_OK;
fail:
aml_gpu_release(gpu);
return status;
}
static zx_driver_ops_t aml_gpu_driver_ops = {
.version = DRIVER_OPS_VERSION,
.bind = aml_gpu_bind,
};
ZIRCON_DRIVER_BEGIN(aml_gpu, aml_gpu_driver_ops, "zircon", "0.1", 6)
BI_ABORT_IF(NE, BIND_PROTOCOL, ZX_PROTOCOL_PDEV),
BI_ABORT_IF(NE, BIND_PLATFORM_DEV_VID, PDEV_VID_AMLOGIC),
BI_ABORT_IF(NE, BIND_PLATFORM_DEV_DID, PDEV_DID_ARM_MALI_INIT),
// we support multiple SOC variants
BI_MATCH_IF(EQ, BIND_PLATFORM_DEV_PID, PDEV_PID_AMLOGIC_S912),
BI_MATCH_IF(EQ, BIND_PLATFORM_DEV_PID, PDEV_PID_AMLOGIC_S905D2),
BI_MATCH_IF(EQ, BIND_PLATFORM_DEV_PID, PDEV_PID_AMLOGIC_T931),
ZIRCON_DRIVER_END(aml_gpu)