Commit 8085ae685477d8c26b7d2afd0f89a0d99adec38e

Con Kolivas 2012-08-11T14:04:55

Further simplify the queue request mechanism.

diff --git a/cgminer.c b/cgminer.c
index f09d57b..525614c 100644
--- a/cgminer.c
+++ b/cgminer.c
@@ -1355,23 +1355,6 @@ static int pending_staged(void)
 	return ret;
 }
 
-static int pool_staged(struct pool *pool)
-{
-	int ret;
-
-	mutex_lock(stgd_lock);
-	ret = pool->staged;
-	mutex_unlock(stgd_lock);
-	return ret;
-}
-
-static int current_staged(void)
-{
-	struct pool *pool = current_pool();
-
-	return pool_staged(pool);
-}
-
 #ifdef HAVE_CURSES
 WINDOW *mainwin, *statuswin, *logwin;
 #endif
@@ -2240,43 +2223,26 @@ static void push_curl_entry(struct curl_ent *ce, struct pool *pool)
 
 /* This is overkill, but at least we'll know accurately how much work is
  * queued to prevent ever being left without work */
-static void inc_queued(struct pool *pool)
+static void inc_queued(void)
 {
 	mutex_lock(stgd_lock);
-	if (likely(pool))
-		pool->queued++;
 	total_queued++;
 	mutex_unlock(stgd_lock);
 }
 
-static void __dec_queued(struct pool *pool)
+static void __dec_queued(void)
 {
-	if (!total_queued)
-		return;
-
-	if (likely(pool))
-		pool->queued--;
-	total_queued--;
+	if (total_queued)
+		total_queued--;
 }
 
-static void dec_queued(struct pool *pool)
+static void dec_queued(void)
 {
 	mutex_lock(stgd_lock);
-	__dec_queued(pool);
+	__dec_queued();
 	mutex_unlock(stgd_lock);
 }
 
-static int current_queued(void)
-{
-	struct pool *pool = current_pool();
-	int ret;
-
-	mutex_lock(stgd_lock);
-	ret = pool->queued;
-	mutex_unlock(stgd_lock);
-	return ret;
-}
-
 static int global_queued(void)
 {
 	int ret;
@@ -2617,7 +2583,6 @@ static void discard_stale(void)
 	HASH_ITER(hh, staged_work, work, tmp) {
 		if (stale_work(work, false)) {
 			HASH_DEL(staged_work, work);
-			work->pool->staged--;
 			discard_work(work);
 			stale++;
 		}
@@ -2789,8 +2754,7 @@ static bool hash_push(struct work *work)
 	mutex_lock(stgd_lock);
 	if (likely(!getq->frozen)) {
 		HASH_ADD_INT(staged_work, id, work);
-		work->pool->staged++;
-		__dec_queued(work->pool);
+		__dec_queued();
 		HASH_SORT(staged_work, tv_sort);
 	} else
 		rc = false;
@@ -3823,34 +3787,25 @@ static bool clone_available(void)
 
 bool queue_request(struct thr_info *thr, bool needed)
 {
-	int cq, cs, ts, tq, maxq;
-	bool lag, ret, qing;
 	struct workio_cmd *wc;
+	bool lag, ret, qing;
+	int ps, ts, maxq;
 
-	inc_queued(NULL);
+	inc_queued();
 
 	maxq = opt_queue + mining_threads;
 	lag = ret = qing = false;
-	cq = current_queued();
-	cs = current_staged();
-	ts = pending_staged();
-	tq = global_queued();
+	ps = pending_staged();
+	ts = total_staged();
 
-	if (needed && cq >= maxq && !ts && !opt_fail_only) {
-		/* If we're queueing work faster than we can stage it, consider
-		 * the system lagging and allow work to be gathered from
-		 * another pool if possible */
-		lag = true;
-	} else {
-		/* Test to make sure we have enough work for pools without rolltime
-		 * and enough original work for pools with rolltime */
-		if (((cs || cq >= opt_queue) && ts >= maxq) ||
-		    ((cs || cq) && tq >= maxq)) {
-			ret = true;
-			goto out;
-		}
+	if (ps >= maxq) {
+		ret = true;
+		goto out;
 	}
 
+	if (needed && !ts && !opt_fail_only)
+		lag = true;
+
 	/* fill out work request message */
 	wc = calloc(1, sizeof(*wc));
 	if (unlikely(!wc)) {
@@ -3873,7 +3828,7 @@ bool queue_request(struct thr_info *thr, bool needed)
 	qing  = ret = true;
 out:
 	if (!qing)
-		dec_queued(NULL);
+		dec_queued();
 	return true;
 }
 
@@ -3889,7 +3844,6 @@ static struct work *hash_pop(const struct timespec *abstime)
 	if (HASH_COUNT(staged_work)) {
 		work = staged_work;
 		HASH_DEL(staged_work, work);
-		work->pool->staged--;
 	}
 	mutex_unlock(stgd_lock);
 
@@ -3952,7 +3906,7 @@ static bool get_work(struct work *work, bool requested, struct thr_info *thr,
 	struct timespec abstime = {0, 0};
 	struct timeval now;
 	struct work *work_heap;
-	int failures = 0, cq;
+	int failures = 0, tq;
 	struct pool *pool;
 
 	/* Tell the watchdog thread this thread is waiting on getwork and
@@ -3965,10 +3919,10 @@ static bool get_work(struct work *work, bool requested, struct thr_info *thr,
 		return true;
 	}
 
-	cq = current_queued();
+	tq = global_queued();
 retry:
 	pool = current_pool();
-	if (!requested || cq < opt_queue) {
+	if (!requested || tq < opt_queue) {
 		if (unlikely(!queue_request(thr, true))) {
 			applog(LOG_WARNING, "Failed to queue_request in get_work");
 			goto out;
@@ -3981,7 +3935,7 @@ retry:
 		goto out;
 	}
 
-	if (!pool->lagging && requested && !newreq && !pool_staged(pool) && cq >= mining_threads + opt_queue) {
+	if (!pool->lagging && requested && !newreq && !total_staged() && pending_staged() >= mining_threads + opt_queue) {
 		struct cgpu_info *cgpu = thr->cgpu;
 		bool stalled = true;
 		int i;
diff --git a/miner.h b/miner.h
index 5afa071..dc94b5a 100644
--- a/miner.h
+++ b/miner.h
@@ -720,8 +720,6 @@ struct pool {
 	int accepted, rejected;
 	int seq_rejects;
 	int solved;
-	int queued;
-	int staged;
 
 	bool submit_fail;
 	bool idle;