/* * Copyright (c) 2018-2020, Andreas Kling * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include #include #include #include #include #include #include #include #include namespace Kernel { InlineLinkedList& all_inodes() { static InlineLinkedList* list; if (!list) list = new InlineLinkedList; return *list; } void Inode::sync() { NonnullRefPtrVector inodes; { InterruptDisabler disabler; for (auto& inode : all_inodes()) { if (inode.is_metadata_dirty()) inodes.append(inode); } } for (auto& inode : inodes) { ASSERT(inode.is_metadata_dirty()); inode.flush_metadata(); } } ByteBuffer Inode::read_entire(FileDescription* descriptor) const { size_t initial_size = metadata().size ? metadata().size : 4096; StringBuilder builder(initial_size); ssize_t nread; u8 buffer[4096]; off_t offset = 0; for (;;) { nread = read_bytes(offset, sizeof(buffer), buffer, descriptor); ASSERT(nread <= (ssize_t)sizeof(buffer)); if (nread <= 0) break; builder.append((const char*)buffer, nread); offset += nread; if (nread < (ssize_t)sizeof(buffer)) break; } if (nread < 0) { klog() << "Inode::read_entire: ERROR: " << nread; return nullptr; } return builder.to_byte_buffer(); } KResultOr> Inode::resolve_as_link(Custody& base, RefPtr* out_parent, int options, int symlink_recursion_level) const { // The default implementation simply treats the stored // contents as a path and resolves that. That is, it // behaves exactly how you would expect a symlink to work. auto contents = read_entire(); if (!contents) { if (out_parent) *out_parent = nullptr; return KResult(-ENOENT); } auto path = StringView(contents.data(), contents.size()); return VFS::the().resolve_path(path, base, out_parent, options, symlink_recursion_level); } Inode::Inode(FS& fs, unsigned index) : m_fs(fs) , m_index(index) { all_inodes().append(this); } Inode::~Inode() { all_inodes().remove(this); } void Inode::will_be_destroyed() { if (m_metadata_dirty) flush_metadata(); } void Inode::inode_contents_changed(off_t offset, ssize_t size, const u8* data) { if (m_shared_vmobject) m_shared_vmobject->inode_contents_changed({}, offset, size, data); } void Inode::inode_size_changed(size_t old_size, size_t new_size) { if (m_shared_vmobject) m_shared_vmobject->inode_size_changed({}, old_size, new_size); } int Inode::set_atime(time_t) { return -ENOTIMPL; } int Inode::set_ctime(time_t) { return -ENOTIMPL; } int Inode::set_mtime(time_t) { return -ENOTIMPL; } KResult Inode::increment_link_count() { return KResult(-ENOTIMPL); } KResult Inode::decrement_link_count() { return KResult(-ENOTIMPL); } void Inode::set_shared_vmobject(SharedInodeVMObject& vmobject) { m_shared_vmobject = vmobject.make_weak_ptr(); } bool Inode::bind_socket(LocalSocket& socket) { LOCKER(m_lock); if (m_socket) return false; m_socket = socket; return true; } bool Inode::unbind_socket() { LOCKER(m_lock); if (!m_socket) return false; m_socket = nullptr; return true; } void Inode::register_watcher(Badge, InodeWatcher& watcher) { LOCKER(m_lock); ASSERT(!m_watchers.contains(&watcher)); m_watchers.set(&watcher); } void Inode::unregister_watcher(Badge, InodeWatcher& watcher) { LOCKER(m_lock); ASSERT(m_watchers.contains(&watcher)); m_watchers.remove(&watcher); } void Inode::set_metadata_dirty(bool metadata_dirty) { if (m_metadata_dirty == metadata_dirty) return; m_metadata_dirty = metadata_dirty; if (m_metadata_dirty) { // FIXME: Maybe we should hook into modification events somewhere else, I'm not sure where. // We don't always end up on this particular code path, for instance when writing to an ext2fs file. LOCKER(m_lock); for (auto& watcher : m_watchers) { watcher->notify_inode_event({}, InodeWatcher::Event::Type::Modified); } } } KResult Inode::prepare_to_write_data() { // FIXME: It's a poor design that filesystems are expected to call this before writing out data. // We should funnel everything through an interface at the VFS layer so this can happen from a single place. LOCKER(m_lock); if (fs().is_readonly()) return KResult(-EROFS); auto metadata = this->metadata(); if (metadata.is_setuid() || metadata.is_setgid()) { dbg() << "Inode::prepare_to_write_data(): Stripping SUID/SGID bits from " << identifier(); return chmod(metadata.mode & ~(04000 | 02000)); } return KSuccess; } }