svn commit: r241249 - stable/9/sys/kern
Alexander Motin
mav at FreeBSD.org
Sat Oct 6 12:58:58 UTC 2012
Author: mav
Date: Sat Oct 6 12:58:56 2012
New Revision: 241249
URL: http://svn.freebsd.org/changeset/base/241249
Log:
MFC r239185, r239196:
Some minor tunings/cleanups inspired by bde@ after previous commits:
- remove extra dynamic variable initializations;
- restore (4BSD) and implement (ULE) hogticks variable setting;
- make sched_rr_interval() more tolerant to options;
- restore (4BSD) and implement (ULE) kern.sched.quantum sysctl, a more
user-friendly wrapper for sched_slice;
- tune some sysctl descriptions;
- make some style fixes.
Modified:
stable/9/sys/kern/sched_4bsd.c
stable/9/sys/kern/sched_ule.c
Directory Properties:
stable/9/sys/ (props changed)
Modified: stable/9/sys/kern/sched_4bsd.c
==============================================================================
--- stable/9/sys/kern/sched_4bsd.c Sat Oct 6 12:51:16 2012 (r241248)
+++ stable/9/sys/kern/sched_4bsd.c Sat Oct 6 12:58:56 2012 (r241249)
@@ -119,9 +119,9 @@ struct td_sched {
static struct td_sched td_sched0;
struct mtx sched_lock;
-static int realstathz; /* stathz is sometimes 0 and run off of hz. */
+static int realstathz = 127; /* stathz is sometimes 0 and run off of hz. */
static int sched_tdcnt; /* Total runnable threads in the system. */
-static int sched_slice = 1; /* Thread run time before rescheduling. */
+static int sched_slice = 12; /* Thread run time before rescheduling. */
static void setup_runqs(void);
static void schedcpu(void);
@@ -185,12 +185,33 @@ setup_runqs(void)
runq_init(&runq);
}
+static int
+sysctl_kern_quantum(SYSCTL_HANDLER_ARGS)
+{
+ int error, new_val, period;
+
+ period = 1000000 / realstathz;
+ new_val = period * sched_slice;
+ error = sysctl_handle_int(oidp, &new_val, 0, req);
+ if (error != 0 || req->newptr == NULL)
+ return (error);
+ if (new_val <= 0)
+ return (EINVAL);
+ sched_slice = imax(1, (new_val + period / 2) / period);
+ hogticks = imax(1, (2 * hz * sched_slice + realstathz / 2) /
+ realstathz);
+ return (0);
+}
+
SYSCTL_NODE(_kern, OID_AUTO, sched, CTLFLAG_RD, 0, "Scheduler");
SYSCTL_STRING(_kern_sched, OID_AUTO, name, CTLFLAG_RD, "4BSD", 0,
"Scheduler name");
+SYSCTL_PROC(_kern_sched, OID_AUTO, quantum, CTLTYPE_INT | CTLFLAG_RW,
+ NULL, 0, sysctl_kern_quantum, "I",
+ "Quantum for timeshare threads in microseconds");
SYSCTL_INT(_kern_sched, OID_AUTO, slice, CTLFLAG_RW, &sched_slice, 0,
- "Slice size for timeshare threads");
+ "Quantum for timeshare threads in stathz ticks");
#ifdef SMP
/* Enable forwarding of wakeups to all other cpus */
SYSCTL_NODE(_kern_sched, OID_AUTO, ipiwakeup, CTLFLAG_RD, NULL, "Kernel SMP");
@@ -628,21 +649,15 @@ resetpriority_thread(struct thread *td)
static void
sched_setup(void *dummy)
{
- setup_runqs();
- /*
- * To avoid divide-by-zero, we set realstathz a dummy value
- * in case which sched_clock() called before sched_initticks().
- */
- realstathz = hz;
- sched_slice = realstathz / 10; /* ~100ms */
+ setup_runqs();
/* Account for thread0. */
sched_load_add();
}
/*
- * This routine determines the sched_slice after stathz and hz are setup.
+ * This routine determines time constants after stathz and hz are setup.
*/
static void
sched_initticks(void *dummy)
@@ -650,6 +665,8 @@ sched_initticks(void *dummy)
realstathz = stathz ? stathz : hz;
sched_slice = realstathz / 10; /* ~100ms */
+ hogticks = imax(1, (2 * hz * sched_slice + realstathz / 2) /
+ realstathz);
}
/* External interfaces start here */
@@ -688,7 +705,7 @@ sched_rr_interval(void)
{
/* Convert sched_slice from stathz to hz. */
- return (hz / (realstathz / sched_slice));
+ return (imax(1, (sched_slice * hz + realstathz / 2) / realstathz));
}
/*
@@ -723,9 +740,9 @@ sched_clock(struct thread *td)
/*
* Force a context switch if the current thread has used up a full
- * quantum (default quantum is 100ms).
+ * time slice (default is 100ms).
*/
- if (!TD_IS_IDLETHREAD(td) && (--ts->ts_slice <= 0)) {
+ if (!TD_IS_IDLETHREAD(td) && --ts->ts_slice <= 0) {
ts->ts_slice = sched_slice;
td->td_flags |= TDF_NEEDRESCHED | TDF_SLICEEND;
}
Modified: stable/9/sys/kern/sched_ule.c
==============================================================================
--- stable/9/sys/kern/sched_ule.c Sat Oct 6 12:51:16 2012 (r241248)
+++ stable/9/sys/kern/sched_ule.c Sat Oct 6 12:58:56 2012 (r241249)
@@ -201,9 +201,9 @@ static struct td_sched td_sched0;
* preempt_thresh: Priority threshold for preemption and remote IPIs.
*/
static int sched_interact = SCHED_INTERACT_THRESH;
-static int realstathz;
-static int tickincr;
-static int sched_slice = 1;
+static int realstathz = 127;
+static int tickincr = 8 << SCHED_TICK_SHIFT;;
+static int sched_slice = 12;
#ifdef PREEMPTION
#ifdef FULL_PREEMPTION
static int preempt_thresh = PRI_MAX_IDLE;
@@ -1363,13 +1363,6 @@ sched_setup(void *dummy)
#else
tdq_setup(tdq);
#endif
- /*
- * To avoid divide-by-zero, we set realstathz a dummy value
- * in case which sched_clock() called before sched_initticks().
- */
- realstathz = hz;
- sched_slice = (realstathz/10); /* ~100ms */
- tickincr = 1 << SCHED_TICK_SHIFT;
/* Add thread0's load since it's running. */
TDQ_LOCK(tdq);
@@ -1380,7 +1373,7 @@ sched_setup(void *dummy)
}
/*
- * This routine determines the tickincr after stathz and hz are setup.
+ * This routine determines time constants after stathz and hz are setup.
*/
/* ARGSUSED */
static void
@@ -1389,7 +1382,9 @@ sched_initticks(void *dummy)
int incr;
realstathz = stathz ? stathz : hz;
- sched_slice = (realstathz/10); /* ~100ms */
+ sched_slice = realstathz / 10; /* ~100ms */
+ hogticks = imax(1, (2 * hz * sched_slice + realstathz / 2) /
+ realstathz);
/*
* tickincr is shifted out by 10 to avoid rounding errors due to
@@ -1418,7 +1413,7 @@ sched_initticks(void *dummy)
affinity = SCHED_AFFINITY_DEFAULT;
#endif
if (sched_idlespinthresh < 0)
- sched_idlespinthresh = max(16, 2 * hz / realstathz);
+ sched_idlespinthresh = imax(16, 2 * hz / realstathz);
}
@@ -1606,8 +1601,8 @@ int
sched_rr_interval(void)
{
- /* Convert sched_slice to hz */
- return (hz/(realstathz/sched_slice));
+ /* Convert sched_slice from stathz to hz. */
+ return (imax(1, (sched_slice * hz + realstathz / 2) / realstathz));
}
/*
@@ -2231,16 +2226,15 @@ sched_clock(struct thread *td)
sched_interact_update(td);
sched_priority(td);
}
+
/*
- * We used up one time slice.
- */
- if (--ts->ts_slice > 0)
- return;
- /*
- * We're out of time, force a requeue at userret().
+ * Force a context switch if the current thread has used up a full
+ * time slice (default is 100ms).
*/
- ts->ts_slice = sched_slice;
- td->td_flags |= TDF_NEEDRESCHED | TDF_SLICEEND;
+ if (!TD_IS_IDLETHREAD(td) && --ts->ts_slice <= 0) {
+ ts->ts_slice = sched_slice;
+ td->td_flags |= TDF_NEEDRESCHED | TDF_SLICEEND;
+ }
}
/*
@@ -2795,21 +2789,44 @@ sysctl_kern_sched_topology_spec(SYSCTL_H
#endif
+static int
+sysctl_kern_quantum(SYSCTL_HANDLER_ARGS)
+{
+ int error, new_val, period;
+
+ period = 1000000 / realstathz;
+ new_val = period * sched_slice;
+ error = sysctl_handle_int(oidp, &new_val, 0, req);
+ if (error != 0 || req->newptr == NULL)
+ return (error);
+ if (new_val <= 0)
+ return (EINVAL);
+ sched_slice = imax(1, (new_val + period / 2) / period);
+ hogticks = imax(1, (2 * hz * sched_slice + realstathz / 2) /
+ realstathz);
+ return (0);
+}
+
SYSCTL_NODE(_kern, OID_AUTO, sched, CTLFLAG_RW, 0, "Scheduler");
SYSCTL_STRING(_kern_sched, OID_AUTO, name, CTLFLAG_RD, "ULE", 0,
"Scheduler name");
+SYSCTL_PROC(_kern_sched, OID_AUTO, quantum, CTLTYPE_INT | CTLFLAG_RW,
+ NULL, 0, sysctl_kern_quantum, "I",
+ "Quantum for timeshare threads in microseconds");
SYSCTL_INT(_kern_sched, OID_AUTO, slice, CTLFLAG_RW, &sched_slice, 0,
- "Slice size for timeshare threads");
+ "Quantum for timeshare threads in stathz ticks");
SYSCTL_INT(_kern_sched, OID_AUTO, interact, CTLFLAG_RW, &sched_interact, 0,
- "Interactivity score threshold");
-SYSCTL_INT(_kern_sched, OID_AUTO, preempt_thresh, CTLFLAG_RW, &preempt_thresh,
- 0,"Min priority for preemption, lower priorities have greater precedence");
-SYSCTL_INT(_kern_sched, OID_AUTO, static_boost, CTLFLAG_RW, &static_boost,
- 0,"Controls whether static kernel priorities are assigned to sleeping threads.");
-SYSCTL_INT(_kern_sched, OID_AUTO, idlespins, CTLFLAG_RW, &sched_idlespins,
- 0,"Number of times idle will spin waiting for new work.");
-SYSCTL_INT(_kern_sched, OID_AUTO, idlespinthresh, CTLFLAG_RW, &sched_idlespinthresh,
- 0,"Threshold before we will permit idle spinning.");
+ "Interactivity score threshold");
+SYSCTL_INT(_kern_sched, OID_AUTO, preempt_thresh, CTLFLAG_RW,
+ &preempt_thresh, 0,
+ "Maximal (lowest) priority for preemption");
+SYSCTL_INT(_kern_sched, OID_AUTO, static_boost, CTLFLAG_RW, &static_boost, 0,
+ "Assign static kernel priorities to sleeping threads");
+SYSCTL_INT(_kern_sched, OID_AUTO, idlespins, CTLFLAG_RW, &sched_idlespins, 0,
+ "Number of times idle thread will spin waiting for new work");
+SYSCTL_INT(_kern_sched, OID_AUTO, idlespinthresh, CTLFLAG_RW,
+ &sched_idlespinthresh, 0,
+ "Threshold before we will permit idle thread spinning");
#ifdef SMP
SYSCTL_INT(_kern_sched, OID_AUTO, affinity, CTLFLAG_RW, &affinity, 0,
"Number of hz ticks to keep thread affinity for");
@@ -2817,17 +2834,14 @@ SYSCTL_INT(_kern_sched, OID_AUTO, balanc
"Enables the long-term load balancer");
SYSCTL_INT(_kern_sched, OID_AUTO, balance_interval, CTLFLAG_RW,
&balance_interval, 0,
- "Average frequency in stathz ticks to run the long-term balancer");
+ "Average period in stathz ticks to run the long-term balancer");
SYSCTL_INT(_kern_sched, OID_AUTO, steal_idle, CTLFLAG_RW, &steal_idle, 0,
"Attempts to steal work from other cores before idling");
SYSCTL_INT(_kern_sched, OID_AUTO, steal_thresh, CTLFLAG_RW, &steal_thresh, 0,
- "Minimum load on remote cpu before we'll steal");
-
-/* Retrieve SMP topology */
+ "Minimum load on remote CPU before we'll steal");
SYSCTL_PROC(_kern_sched, OID_AUTO, topology_spec, CTLTYPE_STRING |
- CTLFLAG_RD, NULL, 0, sysctl_kern_sched_topology_spec, "A",
+ CTLFLAG_RD, NULL, 0, sysctl_kern_sched_topology_spec, "A",
"XML dump of detected CPU topology");
-
#endif
/* ps compat. All cpu percentages from ULE are weighted. */
More information about the svn-src-stable-9
mailing list