8d1ab57065
8300663: java/util/concurrent/SynchronousQueue/Fairness.java failed with "Error: fair=true i=0 j=1" 8267502: JDK-8246677 caused 16x performance regression in SynchronousQueue Reviewed-by: alanb
393 lines
13 KiB
Java
393 lines
13 KiB
Java
/*
|
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
*
|
|
* This code is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 only, as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
* version 2 for more details (a copy is included in the LICENSE file that
|
|
* accompanied this code).
|
|
*
|
|
* You should have received a copy of the GNU General Public License version
|
|
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*
|
|
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
* or visit www.oracle.com if you need additional information or have any
|
|
* questions.
|
|
*/
|
|
|
|
/*
|
|
* This file is available under and governed by the GNU General Public
|
|
* License version 2 only, as published by the Free Software Foundation.
|
|
* However, the following notice accompanied the original version of this
|
|
* file:
|
|
*
|
|
* Written by Martin Buchholz with assistance from members of JCP
|
|
* JSR-166 Expert Group and released to the public domain, as
|
|
* explained at http://creativecommons.org/publicdomain/zero/1.0/
|
|
*/
|
|
|
|
/*
|
|
* @test
|
|
* @modules java.base/java.util.concurrent:open
|
|
* @run testng WhiteBox
|
|
* @summary White box tests of implementation details
|
|
*/
|
|
|
|
import static org.testng.Assert.*;
|
|
import org.testng.annotations.DataProvider;
|
|
import org.testng.annotations.Test;
|
|
|
|
import java.io.ByteArrayInputStream;
|
|
import java.io.ByteArrayOutputStream;
|
|
import java.io.ObjectInputStream;
|
|
import java.io.ObjectOutputStream;
|
|
import java.lang.invoke.MethodHandles;
|
|
import java.lang.invoke.VarHandle;
|
|
import java.util.ArrayList;
|
|
import java.util.Iterator;
|
|
import java.util.List;
|
|
import java.util.concurrent.LinkedTransferQueue;
|
|
import java.util.concurrent.ThreadLocalRandom;
|
|
import java.util.concurrent.TimeUnit;
|
|
import static java.util.stream.Collectors.toList;
|
|
import java.util.function.Consumer;
|
|
|
|
@Test
|
|
public class WhiteBox {
|
|
|
|
final ThreadLocalRandom rnd = ThreadLocalRandom.current();
|
|
final VarHandle HEAD, TAIL, ITEM, NEXT;
|
|
|
|
public WhiteBox() throws Throwable { // throws ReflectiveOperationException {
|
|
try {
|
|
Class<?> qClass = LinkedTransferQueue.class;
|
|
Class<?> nodeClass = Class.forName(qClass.getName() + "$DualNode");
|
|
MethodHandles.Lookup lookup
|
|
= MethodHandles.privateLookupIn(qClass, MethodHandles.lookup());
|
|
HEAD = lookup.findVarHandle(qClass, "head", nodeClass);
|
|
TAIL = lookup.findVarHandle(qClass, "tail", nodeClass);
|
|
NEXT = lookup.findVarHandle(nodeClass, "next", nodeClass);
|
|
ITEM = lookup.findVarHandle(nodeClass, "item", Object.class);
|
|
} catch (Throwable ex) {
|
|
ex.printStackTrace();
|
|
throw ex;
|
|
}
|
|
}
|
|
|
|
Object head(LinkedTransferQueue q) { return HEAD.getVolatile(q); }
|
|
Object tail(LinkedTransferQueue q) { return TAIL.getVolatile(q); }
|
|
Object item(Object node) { return ITEM.getVolatile(node); }
|
|
Object next(Object node) { return NEXT.getVolatile(node); }
|
|
|
|
/*
|
|
* Modified for jdk22: Accommodate lazy initialization, so counts
|
|
* may vary by 1, and some nodes become headers vs unlinked,
|
|
* compared to previous versions.
|
|
*/
|
|
|
|
static void checkCount(int val, int expect) {
|
|
assertTrue(val == expect || val == expect - 1);
|
|
}
|
|
|
|
int nodeCount(LinkedTransferQueue q) {
|
|
int i = 0;
|
|
for (Object p = head(q); p != null; ) {
|
|
i++;
|
|
if (p == (p = next(p))) p = head(q);
|
|
}
|
|
return i;
|
|
}
|
|
|
|
int tailCount(LinkedTransferQueue q) {
|
|
int i = 0;
|
|
for (Object p = tail(q); p != null; ) {
|
|
i++;
|
|
if (p == (p = next(p))) p = head(q);
|
|
}
|
|
return i;
|
|
}
|
|
|
|
Object findNode(LinkedTransferQueue q, Object e) {
|
|
for (Object p = head(q); p != null; ) {
|
|
if (item(p) != null && e.equals(item(p)))
|
|
return p;
|
|
if (p == (p = next(p))) p = head(q);
|
|
}
|
|
throw new AssertionError("not found");
|
|
}
|
|
|
|
Iterator iteratorAt(LinkedTransferQueue q, Object e) {
|
|
for (Iterator it = q.iterator(); it.hasNext(); )
|
|
if (it.next().equals(e))
|
|
return it;
|
|
throw new AssertionError("not found");
|
|
}
|
|
|
|
void assertIsSelfLinked(Object node) {
|
|
assertSame(next(node), node);
|
|
assertNull(item(node));
|
|
}
|
|
void assertIsNotSelfLinked(Object node) {
|
|
assertNotSame(node, next(node));
|
|
}
|
|
|
|
@Test
|
|
public void addRemove() {
|
|
LinkedTransferQueue q = new LinkedTransferQueue();
|
|
assertInvariants(q);
|
|
if (head(q) != null) {
|
|
assertNull(next(head(q)));
|
|
assertNull(item(head(q)));
|
|
}
|
|
q.add(1);
|
|
checkCount(nodeCount(q), 2);
|
|
assertInvariants(q);
|
|
q.remove(1);
|
|
checkCount(nodeCount(q), 1);
|
|
assertInvariants(q);
|
|
}
|
|
|
|
/**
|
|
* Traversal actions that visit every node and do nothing, but
|
|
* have side effect of squeezing out dead nodes.
|
|
*/
|
|
@DataProvider
|
|
public Object[][] traversalActions() {
|
|
return List.<Consumer<LinkedTransferQueue>>of(
|
|
q -> q.forEach(e -> {}),
|
|
q -> assertFalse(q.contains(new Object())),
|
|
q -> assertFalse(q.remove(new Object())),
|
|
q -> q.spliterator().forEachRemaining(e -> {}),
|
|
q -> q.stream().collect(toList()),
|
|
q -> assertFalse(q.removeIf(e -> false)),
|
|
q -> assertFalse(q.removeAll(List.of())))
|
|
.stream().map(x -> new Object[]{ x }).toArray(Object[][]::new);
|
|
}
|
|
|
|
@Test(dataProvider = "traversalActions")
|
|
public void traversalOperationsCollapseLeadingNodes(
|
|
Consumer<LinkedTransferQueue> traversalAction) {
|
|
LinkedTransferQueue q = new LinkedTransferQueue();
|
|
Object oldHead;
|
|
int n = 1 + rnd.nextInt(5);
|
|
for (int i = 0; i < n; i++) q.add(i);
|
|
checkCount(nodeCount(q), n + 1);
|
|
oldHead = head(q);
|
|
traversalAction.accept(q);
|
|
assertInvariants(q);
|
|
checkCount(nodeCount(q), n);
|
|
// assertIsSelfLinked(oldHead);
|
|
}
|
|
|
|
@Test(dataProvider = "traversalActions")
|
|
public void traversalOperationsCollapseInteriorNodes(
|
|
Consumer<LinkedTransferQueue> traversalAction) {
|
|
LinkedTransferQueue q = new LinkedTransferQueue();
|
|
int n = 6;
|
|
for (int i = 0; i < n; i++) q.add(i);
|
|
|
|
// We must be quite devious to reliably create an interior dead node
|
|
Object p0 = findNode(q, 0);
|
|
Object p1 = findNode(q, 1);
|
|
Object p2 = findNode(q, 2);
|
|
Object p3 = findNode(q, 3);
|
|
Object p4 = findNode(q, 4);
|
|
Object p5 = findNode(q, 5);
|
|
|
|
Iterator it1 = iteratorAt(q, 1);
|
|
Iterator it2 = iteratorAt(q, 2);
|
|
|
|
it2.remove(); // causes it2's ancestor to advance to 1
|
|
assertSame(next(p1), p3);
|
|
assertSame(next(p2), p3);
|
|
assertNull(item(p2));
|
|
it1.remove(); // removes it2's ancestor
|
|
assertSame(next(p0), p3);
|
|
assertSame(next(p1), p3);
|
|
assertSame(next(p2), p3);
|
|
assertNull(item(p1));
|
|
assertEquals(it2.next(), 3);
|
|
it2.remove(); // it2's ancestor can't unlink
|
|
|
|
assertSame(next(p0), p3); // p3 is now interior dead node
|
|
assertSame(next(p1), p4); // it2 uselessly CASed p1.next
|
|
assertSame(next(p2), p3);
|
|
assertSame(next(p3), p4);
|
|
assertInvariants(q);
|
|
|
|
int c = nodeCount(q);
|
|
traversalAction.accept(q);
|
|
checkCount(nodeCount(q), c - 1);
|
|
|
|
assertSame(next(p0), p4);
|
|
assertSame(next(p1), p4);
|
|
assertSame(next(p2), p3);
|
|
assertSame(next(p3), p4);
|
|
assertInvariants(q);
|
|
|
|
// trailing nodes are not unlinked
|
|
Iterator it5 = iteratorAt(q, 5); it5.remove();
|
|
traversalAction.accept(q);
|
|
assertSame(next(p4), p5);
|
|
assertNull(next(p5));
|
|
checkCount(nodeCount(q), c - 1);
|
|
}
|
|
|
|
/**
|
|
* Checks that traversal operations collapse a random pattern of
|
|
* dead nodes as could normally only occur with a race.
|
|
*/
|
|
@Test(dataProvider = "traversalActions")
|
|
public void traversalOperationsCollapseRandomNodes(
|
|
Consumer<LinkedTransferQueue> traversalAction) {
|
|
LinkedTransferQueue q = new LinkedTransferQueue();
|
|
int n = 1 + rnd.nextInt(6);
|
|
for (int i = 0; i < n; i++) q.add(i);
|
|
ArrayList nulledOut = new ArrayList();
|
|
for (Object p = head(q); p != null; p = next(p))
|
|
if (rnd.nextBoolean()) {
|
|
nulledOut.add(item(p));
|
|
ITEM.setVolatile(p, null);
|
|
}
|
|
traversalAction.accept(q);
|
|
int c = nodeCount(q);
|
|
checkCount(c - (q.contains(n - 1) ? 0 : 1), q.size() + 1);
|
|
for (int i = 0; i < n; i++)
|
|
assertTrue(nulledOut.contains(i) ^ q.contains(i));
|
|
}
|
|
|
|
/**
|
|
* Traversal actions that remove every element, and are also
|
|
* expected to squeeze out dead nodes.
|
|
*/
|
|
@DataProvider
|
|
public Object[][] bulkRemovalActions() {
|
|
return List.<Consumer<LinkedTransferQueue>>of(
|
|
q -> q.clear(),
|
|
q -> assertTrue(q.removeIf(e -> true)),
|
|
q -> assertTrue(q.retainAll(List.of())))
|
|
.stream().map(x -> new Object[]{ x }).toArray(Object[][]::new);
|
|
}
|
|
|
|
@Test(dataProvider = "bulkRemovalActions")
|
|
public void bulkRemovalOperationsCollapseNodes(
|
|
Consumer<LinkedTransferQueue> bulkRemovalAction) {
|
|
LinkedTransferQueue q = new LinkedTransferQueue();
|
|
int n = 1 + rnd.nextInt(5);
|
|
for (int i = 0; i < n; i++) q.add(i);
|
|
bulkRemovalAction.accept(q);
|
|
checkCount(nodeCount(q), 1);
|
|
assertInvariants(q);
|
|
}
|
|
|
|
/**
|
|
* Actions that remove the first element, and are expected to
|
|
* leave at most one slack dead node at head.
|
|
*/
|
|
@DataProvider
|
|
public Object[][] pollActions() {
|
|
return List.<Consumer<LinkedTransferQueue>>of(
|
|
q -> assertNotNull(q.poll()),
|
|
q -> { try { assertNotNull(q.poll(1L, TimeUnit.DAYS)); }
|
|
catch (Throwable x) { throw new AssertionError(x); }},
|
|
q -> { try { assertNotNull(q.take()); }
|
|
catch (Throwable x) { throw new AssertionError(x); }},
|
|
q -> assertNotNull(q.remove()))
|
|
.stream().map(x -> new Object[]{ x }).toArray(Object[][]::new);
|
|
}
|
|
|
|
@Test(dataProvider = "pollActions")
|
|
public void pollActionsOneNodeSlack(
|
|
Consumer<LinkedTransferQueue> pollAction) {
|
|
LinkedTransferQueue q = new LinkedTransferQueue();
|
|
int n = 1 + rnd.nextInt(5);
|
|
for (int i = 0; i < n; i++) q.add(i);
|
|
checkCount(nodeCount(q), n + 1);
|
|
for (int i = 0; i < n; i++) {
|
|
int c = nodeCount(q);
|
|
boolean slack = item(head(q)) == null;
|
|
if (slack) assertNotNull(item(next(head(q))));
|
|
pollAction.accept(q);
|
|
checkCount(nodeCount(q), q.isEmpty() ? 1 : c - (slack ? 2 : 0));
|
|
}
|
|
assertInvariants(q);
|
|
}
|
|
|
|
/**
|
|
* Actions that append an element, and are expected to
|
|
* leave at most one slack node at tail.
|
|
*/
|
|
@DataProvider
|
|
public Object[][] addActions() {
|
|
return List.<Consumer<LinkedTransferQueue>>of(
|
|
q -> q.add(1),
|
|
q -> q.offer(1))
|
|
.stream().map(x -> new Object[]{ x }).toArray(Object[][]::new);
|
|
}
|
|
|
|
@Test(dataProvider = "addActions")
|
|
public void addActionsOneNodeSlack(
|
|
Consumer<LinkedTransferQueue> addAction) {
|
|
LinkedTransferQueue q = new LinkedTransferQueue();
|
|
int n = 1 + rnd.nextInt(9);
|
|
for (int i = 0; i < n; i++) {
|
|
boolean empty = (tail(q) == null);
|
|
boolean slack = !empty && (next(tail(q)) != null);
|
|
addAction.accept(q);
|
|
if (slack)
|
|
assertNull(next(tail(q)));
|
|
else if (!empty) {
|
|
assertNotNull(next(tail(q)));
|
|
assertNull(next(next(tail(q))));
|
|
}
|
|
assertInvariants(q);
|
|
}
|
|
}
|
|
|
|
byte[] serialBytes(Object o) {
|
|
try {
|
|
ByteArrayOutputStream bos = new ByteArrayOutputStream();
|
|
ObjectOutputStream oos = new ObjectOutputStream(bos);
|
|
oos.writeObject(o);
|
|
oos.flush();
|
|
oos.close();
|
|
return bos.toByteArray();
|
|
} catch (Exception fail) {
|
|
throw new AssertionError(fail);
|
|
}
|
|
}
|
|
|
|
@SuppressWarnings("unchecked")
|
|
<T> T serialClone(T o) {
|
|
try {
|
|
ObjectInputStream ois = new ObjectInputStream
|
|
(new ByteArrayInputStream(serialBytes(o)));
|
|
T clone = (T) ois.readObject();
|
|
assertNotSame(o, clone);
|
|
assertSame(o.getClass(), clone.getClass());
|
|
return clone;
|
|
} catch (Exception fail) {
|
|
throw new AssertionError(fail);
|
|
}
|
|
}
|
|
|
|
@Test
|
|
public void testSerialization() {
|
|
LinkedTransferQueue q = serialClone(new LinkedTransferQueue());
|
|
assertInvariants(q);
|
|
}
|
|
|
|
/** Checks conditions which should always be true. */
|
|
void assertInvariants(LinkedTransferQueue q) {
|
|
// head is never self-linked (but tail may!)
|
|
Object h;
|
|
if ((h = head(q)) != null)
|
|
assertNotSame(h, next(h));
|
|
}
|
|
}
|