mirror of
https://github.com/SerenityOS/serenity.git
synced 2025-01-24 02:12:09 -05:00
b96e4c1308
This change should make it less possible for race conditions to happen and cause fatal errors when accessing the hardware.
220 lines
9 KiB
C++
220 lines
9 KiB
C++
/*
|
|
* Copyright (c) 2021, Liav A. <liavalb@hotmail.co.il>
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* 1. Redistributions of source code must retain the above copyright notice, this
|
|
* list of conditions and the following disclaimer.
|
|
*
|
|
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
|
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
|
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <Kernel/Storage/ATA.h>
|
|
#include <Kernel/Storage/BMIDEChannel.h>
|
|
#include <Kernel/Storage/IDEController.h>
|
|
#include <Kernel/WorkQueue.h>
|
|
|
|
namespace Kernel {
|
|
|
|
UNMAP_AFTER_INIT NonnullRefPtr<BMIDEChannel> BMIDEChannel::create(const IDEController& ide_controller, IDEChannel::IOAddressGroup io_group, IDEChannel::ChannelType type)
|
|
{
|
|
return adopt(*new BMIDEChannel(ide_controller, io_group, type));
|
|
}
|
|
|
|
UNMAP_AFTER_INIT BMIDEChannel::BMIDEChannel(const IDEController& controller, IDEChannel::IOAddressGroup io_group, IDEChannel::ChannelType type)
|
|
: IDEChannel(controller, io_group, type)
|
|
{
|
|
initialize();
|
|
}
|
|
|
|
UNMAP_AFTER_INIT void BMIDEChannel::initialize()
|
|
{
|
|
// Let's try to set up DMA transfers.
|
|
PCI::enable_bus_mastering(m_parent_controller->pci_address());
|
|
m_prdt_page = MM.allocate_supervisor_physical_page();
|
|
m_dma_buffer_page = MM.allocate_supervisor_physical_page();
|
|
if (m_dma_buffer_page.is_null() || m_prdt_page.is_null())
|
|
return;
|
|
m_prdt_region = MM.allocate_kernel_region(m_prdt_page->paddr(), PAGE_SIZE, "IDE PRDT", Region::Access::Read | Region::Access::Write);
|
|
m_dma_buffer_region = MM.allocate_kernel_region(m_dma_buffer_page->paddr(), PAGE_SIZE, "IDE DMA region", Region::Access::Read | Region::Access::Write);
|
|
prdt().end_of_table = 0x8000;
|
|
}
|
|
|
|
static void print_ide_status(u8 status)
|
|
{
|
|
dbgln("BMIDEChannel: print_ide_status: DRQ={} BSY={}, DRDY={}, DSC={}, DF={}, CORR={}, IDX={}, ERR={}",
|
|
(status & ATA_SR_DRQ) != 0,
|
|
(status & ATA_SR_BSY) != 0,
|
|
(status & ATA_SR_DRDY) != 0,
|
|
(status & ATA_SR_DSC) != 0,
|
|
(status & ATA_SR_DF) != 0,
|
|
(status & ATA_SR_CORR) != 0,
|
|
(status & ATA_SR_IDX) != 0,
|
|
(status & ATA_SR_ERR) != 0);
|
|
}
|
|
|
|
void BMIDEChannel::handle_irq(const RegisterState&)
|
|
{
|
|
u8 status = m_io_group.io_base().offset(ATA_REG_STATUS).in<u8>();
|
|
|
|
m_entropy_source.add_random_event(status);
|
|
|
|
VERIFY(m_io_group.bus_master_base().has_value());
|
|
u8 bstatus = m_io_group.bus_master_base().value().offset(2).in<u8>();
|
|
if (!(bstatus & 0x4)) {
|
|
// interrupt not from this device, ignore
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel: ignore interrupt");
|
|
return;
|
|
}
|
|
|
|
ScopedSpinLock lock(m_request_lock);
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel: interrupt: DRQ={}, BSY={}, DRDY={}",
|
|
(status & ATA_SR_DRQ) != 0,
|
|
(status & ATA_SR_BSY) != 0,
|
|
(status & ATA_SR_DRDY) != 0);
|
|
|
|
if (!m_current_request) {
|
|
dbgln("BMIDEChannel: IRQ but no pending request!");
|
|
return;
|
|
}
|
|
|
|
if (status & ATA_SR_ERR) {
|
|
print_ide_status(status);
|
|
m_device_error = m_io_group.io_base().offset(ATA_REG_ERROR).in<u8>();
|
|
dbgln("BMIDEChannel: Error {:#02x}!", (u8)m_device_error);
|
|
try_disambiguate_error();
|
|
complete_current_request(AsyncDeviceRequest::Failure);
|
|
return;
|
|
}
|
|
m_device_error = 0;
|
|
complete_current_request(AsyncDeviceRequest::Success);
|
|
}
|
|
|
|
void BMIDEChannel::complete_current_request(AsyncDeviceRequest::RequestResult result)
|
|
{
|
|
// NOTE: this may be called from the interrupt handler!
|
|
VERIFY(m_current_request);
|
|
VERIFY(m_request_lock.is_locked());
|
|
|
|
// Now schedule reading back the buffer as soon as we leave the irq handler.
|
|
// This is important so that we can safely write the buffer back,
|
|
// which could cause page faults. Note that this may be called immediately
|
|
// before Processor::deferred_call_queue returns!
|
|
g_io_work->queue([this, result]() {
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel::complete_current_request result: {}", (int)result);
|
|
ScopedSpinLock lock(m_request_lock);
|
|
VERIFY(m_current_request);
|
|
auto current_request = m_current_request;
|
|
m_current_request.clear();
|
|
|
|
if (result == AsyncDeviceRequest::Success) {
|
|
if (current_request->request_type() == AsyncBlockDeviceRequest::Read) {
|
|
if (!current_request->write_to_buffer(current_request->buffer(), m_dma_buffer_region->vaddr().as_ptr(), 512 * current_request->block_count())) {
|
|
lock.unlock();
|
|
current_request->complete(AsyncDeviceRequest::MemoryFault);
|
|
return;
|
|
}
|
|
}
|
|
|
|
// I read somewhere that this may trigger a cache flush so let's do it.
|
|
VERIFY(m_io_group.bus_master_base().has_value());
|
|
m_io_group.bus_master_base().value().offset(2).out<u8>(m_io_group.bus_master_base().value().offset(2).in<u8>() | 0x6);
|
|
}
|
|
|
|
lock.unlock();
|
|
current_request->complete(result);
|
|
});
|
|
}
|
|
|
|
void BMIDEChannel::ata_write_sectors(bool slave_request, u16 capabilities)
|
|
{
|
|
VERIFY(m_lock.is_locked());
|
|
VERIFY(!m_current_request.is_null());
|
|
VERIFY(m_current_request->block_count() <= 256);
|
|
|
|
ScopedSpinLock m_lock(m_request_lock);
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel::ata_write_sectors_with_dma ({} x {})", m_current_request->block_index(), m_current_request->block_count());
|
|
|
|
prdt().offset = m_dma_buffer_page->paddr().get();
|
|
prdt().size = 512 * m_current_request->block_count();
|
|
|
|
if (!m_current_request->read_from_buffer(m_current_request->buffer(), m_dma_buffer_region->vaddr().as_ptr(), 512 * m_current_request->block_count())) {
|
|
complete_current_request(AsyncDeviceRequest::MemoryFault);
|
|
return;
|
|
}
|
|
|
|
VERIFY(prdt().size <= PAGE_SIZE);
|
|
VERIFY(m_io_group.bus_master_base().has_value());
|
|
// Stop bus master
|
|
m_io_group.bus_master_base().value().out<u8>(0);
|
|
|
|
// Write the PRDT location
|
|
m_io_group.bus_master_base().value().offset(4).out<u32>(m_prdt_page->paddr().get());
|
|
|
|
// Turn on "Interrupt" and "Error" flag. The error flag should be cleared by hardware.
|
|
m_io_group.bus_master_base().value().offset(2).out<u8>(m_io_group.bus_master_base().value().offset(2).in<u8>() | 0x6);
|
|
|
|
ata_access(Direction::Write, slave_request, m_current_request->block_index(), m_current_request->block_count(), capabilities);
|
|
|
|
// Start bus master
|
|
m_io_group.bus_master_base().value().out<u8>(0x1);
|
|
}
|
|
|
|
void BMIDEChannel::send_ata_io_command(LBAMode lba_mode, Direction direction) const
|
|
{
|
|
if (lba_mode != LBAMode::FortyEightBit) {
|
|
m_io_group.io_base().offset(ATA_REG_COMMAND).out<u8>(direction == Direction::Read ? ATA_CMD_READ_DMA : ATA_CMD_WRITE_DMA);
|
|
} else {
|
|
m_io_group.io_base().offset(ATA_REG_COMMAND).out<u8>(direction == Direction::Read ? ATA_CMD_READ_DMA_EXT : ATA_CMD_WRITE_DMA_EXT);
|
|
}
|
|
}
|
|
|
|
void BMIDEChannel::ata_read_sectors(bool slave_request, u16 capabilities)
|
|
{
|
|
VERIFY(m_lock.is_locked());
|
|
VERIFY(!m_current_request.is_null());
|
|
VERIFY(m_current_request->block_count() <= 256);
|
|
|
|
ScopedSpinLock m_lock(m_request_lock);
|
|
dbgln_if(PATA_DEBUG, "BMIDEChannel::ata_read_sectors_with_dma ({} x {})", m_current_request->block_index(), m_current_request->block_count());
|
|
|
|
prdt().offset = m_dma_buffer_page->paddr().get();
|
|
prdt().size = 512 * m_current_request->block_count();
|
|
|
|
VERIFY(prdt().size <= PAGE_SIZE);
|
|
|
|
VERIFY(m_io_group.bus_master_base().has_value());
|
|
// Stop bus master
|
|
m_io_group.bus_master_base().value().out<u8>(0);
|
|
|
|
// Write the PRDT location
|
|
m_io_group.bus_master_base().value().offset(4).out(m_prdt_page->paddr().get());
|
|
|
|
// Turn on "Interrupt" and "Error" flag. The error flag should be cleared by hardware.
|
|
m_io_group.bus_master_base().value().offset(2).out<u8>(m_io_group.bus_master_base().value().offset(2).in<u8>() | 0x6);
|
|
|
|
// Set transfer direction
|
|
m_io_group.bus_master_base().value().out<u8>(0x8);
|
|
|
|
ata_access(Direction::Read, slave_request, m_current_request->block_index(), m_current_request->block_count(), capabilities);
|
|
|
|
// Start bus master
|
|
m_io_group.bus_master_base().value().out<u8>(0x9);
|
|
}
|
|
|
|
}
|