59376fb2a7
It's no longer used since the removal of the SKAS3/4 support. Signed-off-by: Tiwei Bie <tiwei.btw@antgroup.com> Signed-off-by: Richard Weinberger <richard@nod.at>
1043 lines
28 KiB
C
1043 lines
28 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (C) 2015 Anton Ivanov (aivanov@{brocade.com,kot-begemot.co.uk})
|
|
* Copyright (C) 2015 Thomas Meyer (thomas@m3y3r.de)
|
|
* Copyright (C) 2012-2014 Cisco Systems
|
|
* Copyright (C) 2000 - 2007 Jeff Dike (jdike@{addtoit,linux.intel}.com)
|
|
* Copyright (C) 2019 Intel Corporation
|
|
*/
|
|
|
|
#include <linux/clockchips.h>
|
|
#include <linux/init.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/jiffies.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/threads.h>
|
|
#include <asm/irq.h>
|
|
#include <asm/param.h>
|
|
#include <kern_util.h>
|
|
#include <os.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/time-internal.h>
|
|
#include <linux/um_timetravel.h>
|
|
#include <shared/init.h>
|
|
|
|
#ifdef CONFIG_UML_TIME_TRAVEL_SUPPORT
|
|
enum time_travel_mode time_travel_mode;
|
|
EXPORT_SYMBOL_GPL(time_travel_mode);
|
|
|
|
static bool time_travel_start_set;
|
|
static unsigned long long time_travel_start;
|
|
static unsigned long long time_travel_time;
|
|
static unsigned long long time_travel_shm_offset;
|
|
static LIST_HEAD(time_travel_events);
|
|
static LIST_HEAD(time_travel_irqs);
|
|
static unsigned long long time_travel_timer_interval;
|
|
static unsigned long long time_travel_next_event;
|
|
static struct time_travel_event time_travel_timer_event;
|
|
static int time_travel_ext_fd = -1;
|
|
static unsigned int time_travel_ext_waiting;
|
|
static bool time_travel_ext_prev_request_valid;
|
|
static unsigned long long time_travel_ext_prev_request;
|
|
static unsigned long long *time_travel_ext_free_until;
|
|
static unsigned long long _time_travel_ext_free_until;
|
|
static u16 time_travel_shm_id;
|
|
static struct um_timetravel_schedshm *time_travel_shm;
|
|
static union um_timetravel_schedshm_client *time_travel_shm_client;
|
|
|
|
static void time_travel_set_time(unsigned long long ns)
|
|
{
|
|
if (unlikely(ns < time_travel_time))
|
|
panic("time-travel: time goes backwards %lld -> %lld\n",
|
|
time_travel_time, ns);
|
|
else if (unlikely(ns >= S64_MAX))
|
|
panic("The system was going to sleep forever, aborting");
|
|
|
|
time_travel_time = ns;
|
|
}
|
|
|
|
enum time_travel_message_handling {
|
|
TTMH_IDLE,
|
|
TTMH_POLL,
|
|
TTMH_READ,
|
|
TTMH_READ_START_ACK,
|
|
};
|
|
|
|
static u64 bc_message;
|
|
int time_travel_should_print_bc_msg;
|
|
|
|
void _time_travel_print_bc_msg(void)
|
|
{
|
|
time_travel_should_print_bc_msg = 0;
|
|
printk(KERN_INFO "time-travel: received broadcast 0x%llx\n", bc_message);
|
|
}
|
|
|
|
static void time_travel_setup_shm(int fd, u16 id)
|
|
{
|
|
u32 len;
|
|
|
|
time_travel_shm = os_mmap_rw_shared(fd, sizeof(*time_travel_shm));
|
|
|
|
if (!time_travel_shm)
|
|
goto out;
|
|
|
|
len = time_travel_shm->len;
|
|
|
|
if (time_travel_shm->version != UM_TIMETRAVEL_SCHEDSHM_VERSION ||
|
|
len < struct_size(time_travel_shm, clients, id + 1)) {
|
|
os_unmap_memory(time_travel_shm, sizeof(*time_travel_shm));
|
|
time_travel_shm = NULL;
|
|
goto out;
|
|
}
|
|
|
|
time_travel_shm = os_mremap_rw_shared(time_travel_shm,
|
|
sizeof(*time_travel_shm),
|
|
len);
|
|
if (!time_travel_shm)
|
|
goto out;
|
|
|
|
time_travel_shm_offset = time_travel_shm->current_time;
|
|
time_travel_shm_client = &time_travel_shm->clients[id];
|
|
time_travel_shm_client->capa |= UM_TIMETRAVEL_SCHEDSHM_CAP_TIME_SHARE;
|
|
time_travel_shm_id = id;
|
|
/* always look at that free_until from now on */
|
|
time_travel_ext_free_until = &time_travel_shm->free_until;
|
|
out:
|
|
os_close_file(fd);
|
|
}
|
|
|
|
static void time_travel_handle_message(struct um_timetravel_msg *msg,
|
|
enum time_travel_message_handling mode)
|
|
{
|
|
struct um_timetravel_msg resp = {
|
|
.op = UM_TIMETRAVEL_ACK,
|
|
};
|
|
int ret;
|
|
|
|
/*
|
|
* We can't unlock here, but interrupt signals with a timetravel_handler
|
|
* (see um_request_irq_tt) get to the timetravel_handler anyway.
|
|
*/
|
|
if (mode != TTMH_READ) {
|
|
BUG_ON(mode == TTMH_IDLE && !irqs_disabled());
|
|
|
|
while (os_poll(1, &time_travel_ext_fd) != 0) {
|
|
/* nothing */
|
|
}
|
|
}
|
|
|
|
if (unlikely(mode == TTMH_READ_START_ACK)) {
|
|
int fd[UM_TIMETRAVEL_SHARED_MAX_FDS];
|
|
|
|
ret = os_rcv_fd_msg(time_travel_ext_fd, fd,
|
|
ARRAY_SIZE(fd), msg, sizeof(*msg));
|
|
if (ret == sizeof(*msg)) {
|
|
time_travel_setup_shm(fd[UM_TIMETRAVEL_SHARED_MEMFD],
|
|
msg->time & UM_TIMETRAVEL_START_ACK_ID);
|
|
/* we don't use the logging for now */
|
|
os_close_file(fd[UM_TIMETRAVEL_SHARED_LOGFD]);
|
|
}
|
|
} else {
|
|
ret = os_read_file(time_travel_ext_fd, msg, sizeof(*msg));
|
|
}
|
|
|
|
if (ret == 0)
|
|
panic("time-travel external link is broken\n");
|
|
if (ret != sizeof(*msg))
|
|
panic("invalid time-travel message - %d bytes\n", ret);
|
|
|
|
switch (msg->op) {
|
|
default:
|
|
WARN_ONCE(1, "time-travel: unexpected message %lld\n",
|
|
(unsigned long long)msg->op);
|
|
break;
|
|
case UM_TIMETRAVEL_ACK:
|
|
return;
|
|
case UM_TIMETRAVEL_RUN:
|
|
time_travel_set_time(msg->time);
|
|
if (time_travel_shm) {
|
|
/* no request right now since we're running */
|
|
time_travel_shm_client->flags &=
|
|
~UM_TIMETRAVEL_SCHEDSHM_FLAGS_REQ_RUN;
|
|
/* no ack for shared memory RUN */
|
|
return;
|
|
}
|
|
break;
|
|
case UM_TIMETRAVEL_FREE_UNTIL:
|
|
/* not supposed to get this with shm, but ignore it */
|
|
if (time_travel_shm)
|
|
break;
|
|
time_travel_ext_free_until = &_time_travel_ext_free_until;
|
|
_time_travel_ext_free_until = msg->time;
|
|
break;
|
|
case UM_TIMETRAVEL_BROADCAST:
|
|
bc_message = msg->time;
|
|
time_travel_should_print_bc_msg = 1;
|
|
break;
|
|
}
|
|
|
|
resp.seq = msg->seq;
|
|
os_write_file(time_travel_ext_fd, &resp, sizeof(resp));
|
|
}
|
|
|
|
static u64 time_travel_ext_req(u32 op, u64 time)
|
|
{
|
|
static int seq;
|
|
int mseq = ++seq;
|
|
struct um_timetravel_msg msg = {
|
|
.op = op,
|
|
.time = time,
|
|
.seq = mseq,
|
|
};
|
|
|
|
/*
|
|
* We need to block even the timetravel handlers of SIGIO here and
|
|
* only restore their use when we got the ACK - otherwise we may
|
|
* (will) get interrupted by that, try to queue the IRQ for future
|
|
* processing and thus send another request while we're still waiting
|
|
* for an ACK, but the peer doesn't know we got interrupted and will
|
|
* send the ACKs in the same order as the message, but we'd need to
|
|
* see them in the opposite order ...
|
|
*
|
|
* This wouldn't matter *too* much, but some ACKs carry the
|
|
* current time (for UM_TIMETRAVEL_GET) and getting another
|
|
* ACK without a time would confuse us a lot!
|
|
*
|
|
* The sequence number assignment that happens here lets us
|
|
* debug such message handling issues more easily.
|
|
*/
|
|
block_signals_hard();
|
|
os_write_file(time_travel_ext_fd, &msg, sizeof(msg));
|
|
|
|
/* no ACK expected for WAIT in shared memory mode */
|
|
if (msg.op == UM_TIMETRAVEL_WAIT && time_travel_shm)
|
|
goto done;
|
|
|
|
while (msg.op != UM_TIMETRAVEL_ACK)
|
|
time_travel_handle_message(&msg,
|
|
op == UM_TIMETRAVEL_START ?
|
|
TTMH_READ_START_ACK :
|
|
TTMH_READ);
|
|
|
|
if (msg.seq != mseq)
|
|
panic("time-travel: ACK message has different seqno! op=%d, seq=%d != %d time=%lld\n",
|
|
msg.op, msg.seq, mseq, msg.time);
|
|
|
|
if (op == UM_TIMETRAVEL_GET)
|
|
time_travel_set_time(msg.time);
|
|
done:
|
|
unblock_signals_hard();
|
|
|
|
return msg.time;
|
|
}
|
|
|
|
void __time_travel_wait_readable(int fd)
|
|
{
|
|
int fds[2] = { fd, time_travel_ext_fd };
|
|
int ret;
|
|
|
|
if (time_travel_mode != TT_MODE_EXTERNAL)
|
|
return;
|
|
|
|
while ((ret = os_poll(2, fds))) {
|
|
struct um_timetravel_msg msg;
|
|
|
|
if (ret == 1)
|
|
time_travel_handle_message(&msg, TTMH_READ);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL_GPL(__time_travel_wait_readable);
|
|
|
|
static void time_travel_ext_update_request(unsigned long long time)
|
|
{
|
|
if (time_travel_mode != TT_MODE_EXTERNAL)
|
|
return;
|
|
|
|
/* asked for exactly this time previously */
|
|
if (time_travel_ext_prev_request_valid &&
|
|
time == time_travel_ext_prev_request)
|
|
return;
|
|
|
|
/*
|
|
* if we're running and are allowed to run past the request
|
|
* then we don't need to update it either
|
|
*
|
|
* Note for shm we ignore FREE_UNTIL messages and leave the pointer
|
|
* to shared memory, and for non-shm the offset is 0.
|
|
*/
|
|
if (!time_travel_ext_waiting && time_travel_ext_free_until &&
|
|
time < (*time_travel_ext_free_until - time_travel_shm_offset))
|
|
return;
|
|
|
|
time_travel_ext_prev_request = time;
|
|
time_travel_ext_prev_request_valid = true;
|
|
|
|
if (time_travel_shm) {
|
|
union um_timetravel_schedshm_client *running;
|
|
|
|
running = &time_travel_shm->clients[time_travel_shm->running_id];
|
|
|
|
if (running->capa & UM_TIMETRAVEL_SCHEDSHM_CAP_TIME_SHARE) {
|
|
time_travel_shm_client->flags |=
|
|
UM_TIMETRAVEL_SCHEDSHM_FLAGS_REQ_RUN;
|
|
time += time_travel_shm_offset;
|
|
time_travel_shm_client->req_time = time;
|
|
if (time < time_travel_shm->free_until)
|
|
time_travel_shm->free_until = time;
|
|
return;
|
|
}
|
|
}
|
|
|
|
time_travel_ext_req(UM_TIMETRAVEL_REQUEST, time);
|
|
}
|
|
|
|
void __time_travel_propagate_time(void)
|
|
{
|
|
static unsigned long long last_propagated;
|
|
|
|
if (time_travel_shm) {
|
|
if (time_travel_shm->running_id != time_travel_shm_id)
|
|
panic("time-travel: setting time while not running\n");
|
|
time_travel_shm->current_time = time_travel_time +
|
|
time_travel_shm_offset;
|
|
return;
|
|
}
|
|
|
|
if (last_propagated == time_travel_time)
|
|
return;
|
|
|
|
time_travel_ext_req(UM_TIMETRAVEL_UPDATE, time_travel_time);
|
|
last_propagated = time_travel_time;
|
|
}
|
|
EXPORT_SYMBOL_GPL(__time_travel_propagate_time);
|
|
|
|
/* returns true if we must do a wait to the simtime device */
|
|
static bool time_travel_ext_request(unsigned long long time)
|
|
{
|
|
/*
|
|
* If we received an external sync point ("free until") then we
|
|
* don't have to request/wait for anything until then, unless
|
|
* we're already waiting.
|
|
*
|
|
* Note for shm we ignore FREE_UNTIL messages and leave the pointer
|
|
* to shared memory, and for non-shm the offset is 0.
|
|
*/
|
|
if (!time_travel_ext_waiting && time_travel_ext_free_until &&
|
|
time < (*time_travel_ext_free_until - time_travel_shm_offset))
|
|
return false;
|
|
|
|
time_travel_ext_update_request(time);
|
|
return true;
|
|
}
|
|
|
|
static void time_travel_ext_wait(bool idle)
|
|
{
|
|
struct um_timetravel_msg msg = {
|
|
.op = UM_TIMETRAVEL_ACK,
|
|
};
|
|
|
|
time_travel_ext_prev_request_valid = false;
|
|
if (!time_travel_shm)
|
|
time_travel_ext_free_until = NULL;
|
|
time_travel_ext_waiting++;
|
|
|
|
time_travel_ext_req(UM_TIMETRAVEL_WAIT, -1);
|
|
|
|
/*
|
|
* Here we are deep in the idle loop, so we have to break out of the
|
|
* kernel abstraction in a sense and implement this in terms of the
|
|
* UML system waiting on the VQ interrupt while sleeping, when we get
|
|
* the signal it'll call time_travel_ext_vq_notify_done() completing the
|
|
* call.
|
|
*/
|
|
while (msg.op != UM_TIMETRAVEL_RUN)
|
|
time_travel_handle_message(&msg, idle ? TTMH_IDLE : TTMH_POLL);
|
|
|
|
time_travel_ext_waiting--;
|
|
|
|
/* we might request more stuff while polling - reset when we run */
|
|
time_travel_ext_prev_request_valid = false;
|
|
}
|
|
|
|
static void time_travel_ext_get_time(void)
|
|
{
|
|
if (time_travel_shm)
|
|
time_travel_set_time(time_travel_shm->current_time -
|
|
time_travel_shm_offset);
|
|
else
|
|
time_travel_ext_req(UM_TIMETRAVEL_GET, -1);
|
|
}
|
|
|
|
static void __time_travel_update_time(unsigned long long ns, bool idle)
|
|
{
|
|
if (time_travel_mode == TT_MODE_EXTERNAL && time_travel_ext_request(ns))
|
|
time_travel_ext_wait(idle);
|
|
else
|
|
time_travel_set_time(ns);
|
|
}
|
|
|
|
static struct time_travel_event *time_travel_first_event(void)
|
|
{
|
|
return list_first_entry_or_null(&time_travel_events,
|
|
struct time_travel_event,
|
|
list);
|
|
}
|
|
|
|
static void __time_travel_add_event(struct time_travel_event *e,
|
|
unsigned long long time)
|
|
{
|
|
struct time_travel_event *tmp;
|
|
bool inserted = false;
|
|
unsigned long flags;
|
|
|
|
if (e->pending)
|
|
return;
|
|
|
|
e->pending = true;
|
|
e->time = time;
|
|
|
|
local_irq_save(flags);
|
|
list_for_each_entry(tmp, &time_travel_events, list) {
|
|
/*
|
|
* Add the new entry before one with higher time,
|
|
* or if they're equal and both on stack, because
|
|
* in that case we need to unwind the stack in the
|
|
* right order, and the later event (timer sleep
|
|
* or such) must be dequeued first.
|
|
*/
|
|
if ((tmp->time > e->time) ||
|
|
(tmp->time == e->time && tmp->onstack && e->onstack)) {
|
|
list_add_tail(&e->list, &tmp->list);
|
|
inserted = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!inserted)
|
|
list_add_tail(&e->list, &time_travel_events);
|
|
|
|
tmp = time_travel_first_event();
|
|
time_travel_ext_update_request(tmp->time);
|
|
time_travel_next_event = tmp->time;
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
static void time_travel_add_event(struct time_travel_event *e,
|
|
unsigned long long time)
|
|
{
|
|
if (WARN_ON(!e->fn))
|
|
return;
|
|
|
|
__time_travel_add_event(e, time);
|
|
}
|
|
|
|
void time_travel_add_event_rel(struct time_travel_event *e,
|
|
unsigned long long delay_ns)
|
|
{
|
|
time_travel_add_event(e, time_travel_time + delay_ns);
|
|
}
|
|
|
|
static void time_travel_periodic_timer(struct time_travel_event *e)
|
|
{
|
|
time_travel_add_event(&time_travel_timer_event,
|
|
time_travel_time + time_travel_timer_interval);
|
|
deliver_alarm();
|
|
}
|
|
|
|
void deliver_time_travel_irqs(void)
|
|
{
|
|
struct time_travel_event *e;
|
|
unsigned long flags;
|
|
|
|
/*
|
|
* Don't do anything for most cases. Note that because here we have
|
|
* to disable IRQs (and re-enable later) we'll actually recurse at
|
|
* the end of the function, so this is strictly necessary.
|
|
*/
|
|
if (likely(list_empty(&time_travel_irqs)))
|
|
return;
|
|
|
|
local_irq_save(flags);
|
|
irq_enter();
|
|
while ((e = list_first_entry_or_null(&time_travel_irqs,
|
|
struct time_travel_event,
|
|
list))) {
|
|
list_del(&e->list);
|
|
e->pending = false;
|
|
e->fn(e);
|
|
}
|
|
irq_exit();
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
static void time_travel_deliver_event(struct time_travel_event *e)
|
|
{
|
|
if (e == &time_travel_timer_event) {
|
|
/*
|
|
* deliver_alarm() does the irq_enter/irq_exit
|
|
* by itself, so must handle it specially here
|
|
*/
|
|
e->fn(e);
|
|
} else if (irqs_disabled()) {
|
|
list_add_tail(&e->list, &time_travel_irqs);
|
|
/*
|
|
* set pending again, it was set to false when the
|
|
* event was deleted from the original list, but
|
|
* now it's still pending until we deliver the IRQ.
|
|
*/
|
|
e->pending = true;
|
|
} else {
|
|
unsigned long flags;
|
|
|
|
local_irq_save(flags);
|
|
irq_enter();
|
|
e->fn(e);
|
|
irq_exit();
|
|
local_irq_restore(flags);
|
|
}
|
|
}
|
|
|
|
bool time_travel_del_event(struct time_travel_event *e)
|
|
{
|
|
unsigned long flags;
|
|
|
|
if (!e->pending)
|
|
return false;
|
|
local_irq_save(flags);
|
|
list_del(&e->list);
|
|
e->pending = false;
|
|
local_irq_restore(flags);
|
|
return true;
|
|
}
|
|
|
|
static void time_travel_update_time(unsigned long long next, bool idle)
|
|
{
|
|
struct time_travel_event ne = {
|
|
.onstack = true,
|
|
};
|
|
struct time_travel_event *e;
|
|
bool finished = idle;
|
|
|
|
/* add it without a handler - we deal with that specifically below */
|
|
__time_travel_add_event(&ne, next);
|
|
|
|
do {
|
|
e = time_travel_first_event();
|
|
|
|
BUG_ON(!e);
|
|
__time_travel_update_time(e->time, idle);
|
|
|
|
/* new events may have been inserted while we were waiting */
|
|
if (e == time_travel_first_event()) {
|
|
BUG_ON(!time_travel_del_event(e));
|
|
BUG_ON(time_travel_time != e->time);
|
|
|
|
if (e == &ne) {
|
|
finished = true;
|
|
} else {
|
|
if (e->onstack)
|
|
panic("On-stack event dequeued outside of the stack! time=%lld, event time=%lld, event=%pS\n",
|
|
time_travel_time, e->time, e);
|
|
time_travel_deliver_event(e);
|
|
}
|
|
}
|
|
|
|
e = time_travel_first_event();
|
|
if (e)
|
|
time_travel_ext_update_request(e->time);
|
|
} while (ne.pending && !finished);
|
|
|
|
time_travel_del_event(&ne);
|
|
}
|
|
|
|
static void time_travel_update_time_rel(unsigned long long offs)
|
|
{
|
|
unsigned long flags;
|
|
|
|
/*
|
|
* Disable interrupts before calculating the new time so
|
|
* that a real timer interrupt (signal) can't happen at
|
|
* a bad time e.g. after we read time_travel_time but
|
|
* before we've completed updating the time.
|
|
*/
|
|
local_irq_save(flags);
|
|
time_travel_update_time(time_travel_time + offs, false);
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
void time_travel_ndelay(unsigned long nsec)
|
|
{
|
|
/*
|
|
* Not strictly needed to use _rel() version since this is
|
|
* only used in INFCPU/EXT modes, but it doesn't hurt and
|
|
* is more readable too.
|
|
*/
|
|
time_travel_update_time_rel(nsec);
|
|
}
|
|
EXPORT_SYMBOL(time_travel_ndelay);
|
|
|
|
void time_travel_add_irq_event(struct time_travel_event *e)
|
|
{
|
|
BUG_ON(time_travel_mode != TT_MODE_EXTERNAL);
|
|
|
|
time_travel_ext_get_time();
|
|
/*
|
|
* We could model interrupt latency here, for now just
|
|
* don't have any latency at all and request the exact
|
|
* same time (again) to run the interrupt...
|
|
*/
|
|
time_travel_add_event(e, time_travel_time);
|
|
}
|
|
EXPORT_SYMBOL_GPL(time_travel_add_irq_event);
|
|
|
|
static void time_travel_oneshot_timer(struct time_travel_event *e)
|
|
{
|
|
deliver_alarm();
|
|
}
|
|
|
|
void time_travel_sleep(void)
|
|
{
|
|
/*
|
|
* Wait "forever" (using S64_MAX because there are some potential
|
|
* wrapping issues, especially with the current TT_MODE_EXTERNAL
|
|
* controller application.
|
|
*/
|
|
unsigned long long next = S64_MAX;
|
|
|
|
if (time_travel_mode == TT_MODE_BASIC)
|
|
os_timer_disable();
|
|
|
|
time_travel_update_time(next, true);
|
|
|
|
if (time_travel_mode == TT_MODE_BASIC &&
|
|
time_travel_timer_event.pending) {
|
|
if (time_travel_timer_event.fn == time_travel_periodic_timer) {
|
|
/*
|
|
* This is somewhat wrong - we should get the first
|
|
* one sooner like the os_timer_one_shot() below...
|
|
*/
|
|
os_timer_set_interval(time_travel_timer_interval);
|
|
} else {
|
|
os_timer_one_shot(time_travel_timer_event.time - next);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void time_travel_handle_real_alarm(void)
|
|
{
|
|
time_travel_set_time(time_travel_next_event);
|
|
|
|
time_travel_del_event(&time_travel_timer_event);
|
|
|
|
if (time_travel_timer_event.fn == time_travel_periodic_timer)
|
|
time_travel_add_event(&time_travel_timer_event,
|
|
time_travel_time +
|
|
time_travel_timer_interval);
|
|
}
|
|
|
|
static void time_travel_set_interval(unsigned long long interval)
|
|
{
|
|
time_travel_timer_interval = interval;
|
|
}
|
|
|
|
static int time_travel_connect_external(const char *socket)
|
|
{
|
|
const char *sep;
|
|
unsigned long long id = (unsigned long long)-1;
|
|
int rc;
|
|
|
|
if ((sep = strchr(socket, ':'))) {
|
|
char buf[25] = {};
|
|
if (sep - socket > sizeof(buf) - 1)
|
|
goto invalid_number;
|
|
|
|
memcpy(buf, socket, sep - socket);
|
|
if (kstrtoull(buf, 0, &id)) {
|
|
invalid_number:
|
|
panic("time-travel: invalid external ID in string '%s'\n",
|
|
socket);
|
|
return -EINVAL;
|
|
}
|
|
|
|
socket = sep + 1;
|
|
}
|
|
|
|
rc = os_connect_socket(socket);
|
|
if (rc < 0) {
|
|
panic("time-travel: failed to connect to external socket %s\n",
|
|
socket);
|
|
return rc;
|
|
}
|
|
|
|
time_travel_ext_fd = rc;
|
|
|
|
time_travel_ext_req(UM_TIMETRAVEL_START, id);
|
|
|
|
return 1;
|
|
}
|
|
|
|
static void time_travel_set_start(void)
|
|
{
|
|
if (time_travel_start_set)
|
|
return;
|
|
|
|
switch (time_travel_mode) {
|
|
case TT_MODE_EXTERNAL:
|
|
time_travel_start = time_travel_ext_req(UM_TIMETRAVEL_GET_TOD, -1);
|
|
/* controller gave us the *current* time, so adjust by that */
|
|
time_travel_ext_get_time();
|
|
time_travel_start -= time_travel_time;
|
|
break;
|
|
case TT_MODE_INFCPU:
|
|
case TT_MODE_BASIC:
|
|
if (!time_travel_start_set)
|
|
time_travel_start = os_persistent_clock_emulation();
|
|
break;
|
|
case TT_MODE_OFF:
|
|
/* we just read the host clock with os_persistent_clock_emulation() */
|
|
break;
|
|
}
|
|
|
|
time_travel_start_set = true;
|
|
}
|
|
#else /* CONFIG_UML_TIME_TRAVEL_SUPPORT */
|
|
#define time_travel_start_set 0
|
|
#define time_travel_start 0
|
|
#define time_travel_time 0
|
|
#define time_travel_ext_waiting 0
|
|
|
|
static inline void time_travel_update_time(unsigned long long ns, bool idle)
|
|
{
|
|
}
|
|
|
|
static inline void time_travel_update_time_rel(unsigned long long offs)
|
|
{
|
|
}
|
|
|
|
static inline void time_travel_handle_real_alarm(void)
|
|
{
|
|
}
|
|
|
|
static void time_travel_set_interval(unsigned long long interval)
|
|
{
|
|
}
|
|
|
|
static inline void time_travel_set_start(void)
|
|
{
|
|
}
|
|
|
|
/* fail link if this actually gets used */
|
|
extern u64 time_travel_ext_req(u32 op, u64 time);
|
|
|
|
/* these are empty macros so the struct/fn need not exist */
|
|
#define time_travel_add_event(e, time) do { } while (0)
|
|
/* externally not usable - redefine here so we can */
|
|
#undef time_travel_del_event
|
|
#define time_travel_del_event(e) do { } while (0)
|
|
#endif
|
|
|
|
void timer_handler(int sig, struct siginfo *unused_si, struct uml_pt_regs *regs)
|
|
{
|
|
unsigned long flags;
|
|
|
|
/*
|
|
* In basic time-travel mode we still get real interrupts
|
|
* (signals) but since we don't read time from the OS, we
|
|
* must update the simulated time here to the expiry when
|
|
* we get a signal.
|
|
* This is not the case in inf-cpu mode, since there we
|
|
* never get any real signals from the OS.
|
|
*/
|
|
if (time_travel_mode == TT_MODE_BASIC)
|
|
time_travel_handle_real_alarm();
|
|
|
|
local_irq_save(flags);
|
|
do_IRQ(TIMER_IRQ, regs);
|
|
local_irq_restore(flags);
|
|
}
|
|
|
|
static int itimer_shutdown(struct clock_event_device *evt)
|
|
{
|
|
if (time_travel_mode != TT_MODE_OFF)
|
|
time_travel_del_event(&time_travel_timer_event);
|
|
|
|
if (time_travel_mode != TT_MODE_INFCPU &&
|
|
time_travel_mode != TT_MODE_EXTERNAL)
|
|
os_timer_disable();
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int itimer_set_periodic(struct clock_event_device *evt)
|
|
{
|
|
unsigned long long interval = NSEC_PER_SEC / HZ;
|
|
|
|
if (time_travel_mode != TT_MODE_OFF) {
|
|
time_travel_del_event(&time_travel_timer_event);
|
|
time_travel_set_event_fn(&time_travel_timer_event,
|
|
time_travel_periodic_timer);
|
|
time_travel_set_interval(interval);
|
|
time_travel_add_event(&time_travel_timer_event,
|
|
time_travel_time + interval);
|
|
}
|
|
|
|
if (time_travel_mode != TT_MODE_INFCPU &&
|
|
time_travel_mode != TT_MODE_EXTERNAL)
|
|
os_timer_set_interval(interval);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int itimer_next_event(unsigned long delta,
|
|
struct clock_event_device *evt)
|
|
{
|
|
delta += 1;
|
|
|
|
if (time_travel_mode != TT_MODE_OFF) {
|
|
time_travel_del_event(&time_travel_timer_event);
|
|
time_travel_set_event_fn(&time_travel_timer_event,
|
|
time_travel_oneshot_timer);
|
|
time_travel_add_event(&time_travel_timer_event,
|
|
time_travel_time + delta);
|
|
}
|
|
|
|
if (time_travel_mode != TT_MODE_INFCPU &&
|
|
time_travel_mode != TT_MODE_EXTERNAL)
|
|
return os_timer_one_shot(delta);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int itimer_one_shot(struct clock_event_device *evt)
|
|
{
|
|
return itimer_next_event(0, evt);
|
|
}
|
|
|
|
static struct clock_event_device timer_clockevent = {
|
|
.name = "posix-timer",
|
|
.rating = 250,
|
|
.cpumask = cpu_possible_mask,
|
|
.features = CLOCK_EVT_FEAT_PERIODIC |
|
|
CLOCK_EVT_FEAT_ONESHOT,
|
|
.set_state_shutdown = itimer_shutdown,
|
|
.set_state_periodic = itimer_set_periodic,
|
|
.set_state_oneshot = itimer_one_shot,
|
|
.set_next_event = itimer_next_event,
|
|
.shift = 0,
|
|
.max_delta_ns = 0xffffffff,
|
|
.max_delta_ticks = 0xffffffff,
|
|
.min_delta_ns = TIMER_MIN_DELTA,
|
|
.min_delta_ticks = TIMER_MIN_DELTA, // microsecond resolution should be enough for anyone, same as 640K RAM
|
|
.irq = 0,
|
|
.mult = 1,
|
|
};
|
|
|
|
static irqreturn_t um_timer(int irq, void *dev)
|
|
{
|
|
if (get_current()->mm != NULL)
|
|
{
|
|
/* userspace - relay signal, results in correct userspace timers */
|
|
os_alarm_process(get_current()->mm->context.id.pid);
|
|
}
|
|
|
|
(*timer_clockevent.event_handler)(&timer_clockevent);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static u64 timer_read(struct clocksource *cs)
|
|
{
|
|
if (time_travel_mode != TT_MODE_OFF) {
|
|
/*
|
|
* We make reading the timer cost a bit so that we don't get
|
|
* stuck in loops that expect time to move more than the
|
|
* exact requested sleep amount, e.g. python's socket server,
|
|
* see https://bugs.python.org/issue37026.
|
|
*
|
|
* However, don't do that when we're in interrupt or such as
|
|
* then we might recurse into our own processing, and get to
|
|
* even more waiting, and that's not good - it messes up the
|
|
* "what do I do next" and onstack event we use to know when
|
|
* to return from time_travel_update_time().
|
|
*/
|
|
if (!irqs_disabled() && !in_interrupt() && !in_softirq() &&
|
|
!time_travel_ext_waiting)
|
|
time_travel_update_time_rel(TIMER_MULTIPLIER);
|
|
return time_travel_time / TIMER_MULTIPLIER;
|
|
}
|
|
|
|
return os_nsecs() / TIMER_MULTIPLIER;
|
|
}
|
|
|
|
static struct clocksource timer_clocksource = {
|
|
.name = "timer",
|
|
.rating = 300,
|
|
.read = timer_read,
|
|
.mask = CLOCKSOURCE_MASK(64),
|
|
.flags = CLOCK_SOURCE_IS_CONTINUOUS,
|
|
};
|
|
|
|
static void __init um_timer_setup(void)
|
|
{
|
|
int err;
|
|
|
|
err = request_irq(TIMER_IRQ, um_timer, IRQF_TIMER, "hr timer", NULL);
|
|
if (err != 0)
|
|
printk(KERN_ERR "register_timer : request_irq failed - "
|
|
"errno = %d\n", -err);
|
|
|
|
err = os_timer_create();
|
|
if (err != 0) {
|
|
printk(KERN_ERR "creation of timer failed - errno = %d\n", -err);
|
|
return;
|
|
}
|
|
|
|
err = clocksource_register_hz(&timer_clocksource, NSEC_PER_SEC/TIMER_MULTIPLIER);
|
|
if (err) {
|
|
printk(KERN_ERR "clocksource_register_hz returned %d\n", err);
|
|
return;
|
|
}
|
|
clockevents_register_device(&timer_clockevent);
|
|
}
|
|
|
|
void read_persistent_clock64(struct timespec64 *ts)
|
|
{
|
|
long long nsecs;
|
|
|
|
time_travel_set_start();
|
|
|
|
if (time_travel_mode != TT_MODE_OFF)
|
|
nsecs = time_travel_start + time_travel_time;
|
|
else
|
|
nsecs = os_persistent_clock_emulation();
|
|
|
|
set_normalized_timespec64(ts, nsecs / NSEC_PER_SEC,
|
|
nsecs % NSEC_PER_SEC);
|
|
}
|
|
|
|
void __init time_init(void)
|
|
{
|
|
timer_set_signal_handler();
|
|
late_time_init = um_timer_setup;
|
|
}
|
|
|
|
#ifdef CONFIG_UML_TIME_TRAVEL_SUPPORT
|
|
unsigned long calibrate_delay_is_known(void)
|
|
{
|
|
if (time_travel_mode == TT_MODE_INFCPU ||
|
|
time_travel_mode == TT_MODE_EXTERNAL)
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
static int setup_time_travel(char *str)
|
|
{
|
|
if (strcmp(str, "=inf-cpu") == 0) {
|
|
time_travel_mode = TT_MODE_INFCPU;
|
|
timer_clockevent.name = "time-travel-timer-infcpu";
|
|
timer_clocksource.name = "time-travel-clock";
|
|
return 1;
|
|
}
|
|
|
|
if (strncmp(str, "=ext:", 5) == 0) {
|
|
time_travel_mode = TT_MODE_EXTERNAL;
|
|
timer_clockevent.name = "time-travel-timer-external";
|
|
timer_clocksource.name = "time-travel-clock-external";
|
|
return time_travel_connect_external(str + 5);
|
|
}
|
|
|
|
if (!*str) {
|
|
time_travel_mode = TT_MODE_BASIC;
|
|
timer_clockevent.name = "time-travel-timer";
|
|
timer_clocksource.name = "time-travel-clock";
|
|
return 1;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
__setup("time-travel", setup_time_travel);
|
|
__uml_help(setup_time_travel,
|
|
"time-travel\n"
|
|
"This option just enables basic time travel mode, in which the clock/timers\n"
|
|
"inside the UML instance skip forward when there's nothing to do, rather than\n"
|
|
"waiting for real time to elapse. However, instance CPU speed is limited by\n"
|
|
"the real CPU speed, so e.g. a 10ms timer will always fire after ~10ms wall\n"
|
|
"clock (but quicker when there's nothing to do).\n"
|
|
"\n"
|
|
"time-travel=inf-cpu\n"
|
|
"This enables time travel mode with infinite processing power, in which there\n"
|
|
"are no wall clock timers, and any CPU processing happens - as seen from the\n"
|
|
"guest - instantly. This can be useful for accurate simulation regardless of\n"
|
|
"debug overhead, physical CPU speed, etc. but is somewhat dangerous as it can\n"
|
|
"easily lead to getting stuck (e.g. if anything in the system busy loops).\n"
|
|
"\n"
|
|
"time-travel=ext:[ID:]/path/to/socket\n"
|
|
"This enables time travel mode similar to =inf-cpu, except the system will\n"
|
|
"use the given socket to coordinate with a central scheduler, in order to\n"
|
|
"have more than one system simultaneously be on simulated time. The virtio\n"
|
|
"driver code in UML knows about this so you can also simulate networks and\n"
|
|
"devices using it, assuming the device has the right capabilities.\n"
|
|
"The optional ID is a 64-bit integer that's sent to the central scheduler.\n");
|
|
|
|
static int setup_time_travel_start(char *str)
|
|
{
|
|
int err;
|
|
|
|
err = kstrtoull(str, 0, &time_travel_start);
|
|
if (err)
|
|
return err;
|
|
|
|
time_travel_start_set = 1;
|
|
return 1;
|
|
}
|
|
|
|
__setup("time-travel-start=", setup_time_travel_start);
|
|
__uml_help(setup_time_travel_start,
|
|
"time-travel-start=<nanoseconds>\n"
|
|
"Configure the UML instance's wall clock to start at this value rather than\n"
|
|
"the host's wall clock at the time of UML boot.\n");
|
|
static struct kobject *bc_time_kobject;
|
|
|
|
static ssize_t bc_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf)
|
|
{
|
|
return sprintf(buf, "0x%llx", bc_message);
|
|
}
|
|
|
|
static ssize_t bc_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
int ret;
|
|
u64 user_bc_message;
|
|
|
|
ret = kstrtou64(buf, 0, &user_bc_message);
|
|
if (ret)
|
|
return ret;
|
|
|
|
bc_message = user_bc_message;
|
|
|
|
time_travel_ext_req(UM_TIMETRAVEL_BROADCAST, bc_message);
|
|
pr_info("um: time: sent broadcast message: 0x%llx\n", bc_message);
|
|
return count;
|
|
}
|
|
|
|
static struct kobj_attribute bc_attribute = __ATTR(bc-message, 0660, bc_show, bc_store);
|
|
|
|
static int __init um_bc_start(void)
|
|
{
|
|
if (time_travel_mode != TT_MODE_EXTERNAL)
|
|
return 0;
|
|
|
|
bc_time_kobject = kobject_create_and_add("um-ext-time", kernel_kobj);
|
|
if (!bc_time_kobject)
|
|
return 0;
|
|
|
|
if (sysfs_create_file(bc_time_kobject, &bc_attribute.attr))
|
|
pr_debug("failed to create the bc file in /sys/kernel/um_time");
|
|
|
|
return 0;
|
|
}
|
|
late_initcall(um_bc_start);
|
|
#endif
|