7993a62d04
The vectx API, computes the hash for verifying a file as it is read. This avoids the overhead of reading files twice - once to verify, then again to load. For doing an install via loader, avoiding the need to rewind large files is critical. This API is only used for modules, kernel and mdimage as these are the biggest files read by the loader. The reduction in boot time depends on how expensive the I/O is on any given platform. On a fast VM we see 6% improvement. For install via loader the first file to be verified is likely to be the kernel, so some of the prep work (finding manifest etc) done by verify_file() needs to be factored so it can be reused for vectx_open(). For missing or unrecognized fingerprint entries, we fail in vectx_open() unless verifying is disabled. Otherwise fingerprint check happens in vectx_close() and since this API is only used for files which must be verified (VE_MUST) we panic if we get an incorrect hash. Reviewed by: imp,tsoome MFC after: 1 week Sponsored by: Juniper Networks Differential Revision: https://reviews.freebsd.org//D23827
174 lines
4.5 KiB
C
174 lines
4.5 KiB
C
/*-
|
|
* Copyright (c) 1998 Michael Smith <msmith@freebsd.org>
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
/*
|
|
* MD primitives supporting placement of module data
|
|
*
|
|
* XXX should check load address/size against memory top.
|
|
*/
|
|
#include <stand.h>
|
|
|
|
#include "libofw.h"
|
|
|
|
#define READIN_BUF (4 * 1024)
|
|
#define PAGE_SIZE 0x1000
|
|
#define PAGE_MASK 0x0fff
|
|
#define MAPMEM_PAGE_INC 128 /* Half-MB at a time */
|
|
|
|
|
|
#define roundup(x, y) ((((x)+((y)-1))/(y))*(y))
|
|
|
|
static int
|
|
ofw_mapmem(vm_offset_t dest, const size_t len)
|
|
{
|
|
void *destp, *addr;
|
|
size_t dlen;
|
|
size_t resid;
|
|
size_t nlen;
|
|
static vm_offset_t last_dest = 0;
|
|
static size_t last_len = 0;
|
|
|
|
nlen = len;
|
|
/*
|
|
* Check to see if this region fits in a prior mapping.
|
|
* Allocations are generally sequential, so only check
|
|
* the last one.
|
|
*/
|
|
if (dest >= last_dest &&
|
|
(dest + len) <= (last_dest + last_len)) {
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Trim area covered by existing mapping, if any
|
|
*/
|
|
if (dest < (last_dest + last_len) && dest >= last_dest) {
|
|
nlen -= (last_dest + last_len) - dest;
|
|
dest = last_dest + last_len;
|
|
}
|
|
|
|
destp = (void *)(dest & ~PAGE_MASK);
|
|
resid = dest & PAGE_MASK;
|
|
|
|
/*
|
|
* To avoid repeated mappings on small allocations,
|
|
* never map anything less than MAPMEM_PAGE_INC pages at a time
|
|
*/
|
|
if ((nlen + resid) < PAGE_SIZE*MAPMEM_PAGE_INC) {
|
|
dlen = PAGE_SIZE*MAPMEM_PAGE_INC;
|
|
} else
|
|
dlen = roundup(nlen + resid, PAGE_SIZE);
|
|
|
|
if (OF_call_method("claim", memory, 3, 1, destp, dlen, 0, &addr)
|
|
== -1) {
|
|
printf("ofw_mapmem: physical claim failed\n");
|
|
return (ENOMEM);
|
|
}
|
|
|
|
/*
|
|
* We only do virtual memory management when real_mode is false.
|
|
*/
|
|
if (real_mode == 0) {
|
|
if (OF_call_method("claim", mmu, 3, 1, destp, dlen, 0, &addr)
|
|
== -1) {
|
|
printf("ofw_mapmem: virtual claim failed\n");
|
|
return (ENOMEM);
|
|
}
|
|
|
|
if (OF_call_method("map", mmu, 4, 0, destp, destp, dlen, 0)
|
|
== -1) {
|
|
printf("ofw_mapmem: map failed\n");
|
|
return (ENOMEM);
|
|
}
|
|
}
|
|
last_dest = (vm_offset_t) destp;
|
|
last_len = dlen;
|
|
|
|
return (0);
|
|
}
|
|
|
|
ssize_t
|
|
ofw_copyin(const void *src, vm_offset_t dest, const size_t len)
|
|
{
|
|
if (ofw_mapmem(dest, len)) {
|
|
printf("ofw_copyin: map error\n");
|
|
return (0);
|
|
}
|
|
|
|
bcopy(src, (void *)dest, len);
|
|
return(len);
|
|
}
|
|
|
|
ssize_t
|
|
ofw_copyout(const vm_offset_t src, void *dest, const size_t len)
|
|
{
|
|
bcopy((void *)src, dest, len);
|
|
return(len);
|
|
}
|
|
|
|
ssize_t
|
|
ofw_readin(readin_handle_t fd, vm_offset_t dest, const size_t len)
|
|
{
|
|
void *buf;
|
|
size_t resid, chunk, get;
|
|
ssize_t got;
|
|
vm_offset_t p;
|
|
|
|
p = dest;
|
|
|
|
chunk = min(READIN_BUF, len);
|
|
buf = malloc(chunk);
|
|
if (buf == NULL) {
|
|
printf("ofw_readin: buf malloc failed\n");
|
|
return(0);
|
|
}
|
|
|
|
if (ofw_mapmem(dest, len)) {
|
|
printf("ofw_readin: map error\n");
|
|
free(buf);
|
|
return (0);
|
|
}
|
|
|
|
for (resid = len; resid > 0; resid -= got, p += got) {
|
|
get = min(chunk, resid);
|
|
got = VECTX_READ(fd, buf, get);
|
|
|
|
if (got <= 0) {
|
|
if (got < 0)
|
|
printf("ofw_readin: read failed\n");
|
|
break;
|
|
}
|
|
|
|
bcopy(buf, (void *)p, got);
|
|
}
|
|
|
|
free(buf);
|
|
return(len - resid);
|
|
}
|