Lines Matching full:c1
179 Coroutine *c1; in test_no_dangling_access() local
184 c1 = qemu_coroutine_create(c1_fn, c2); in test_no_dangling_access()
186 qemu_coroutine_enter(c1); in test_no_dangling_access()
188 /* c1 shouldn't be used any more now; make sure we segfault if it is */ in test_no_dangling_access()
189 tmp = *c1; in test_no_dangling_access()
190 memset(c1, 0xff, sizeof(Coroutine)); in test_no_dangling_access()
194 *c1 = tmp; in test_no_dangling_access()
226 Coroutine *c1 = qemu_coroutine_create(entry, opaque); in do_test_co_mutex() local
230 qemu_coroutine_enter(c1); in do_test_co_mutex()
234 /* Unlock queues c2. It is then started automatically when c1 yields or in do_test_co_mutex()
237 qemu_coroutine_enter(c1); in do_test_co_mutex()
271 * | c1 | c2 |
308 Coroutine *c1, *c2; in test_co_rwlock_upgrade() local
311 c1 = qemu_coroutine_create(rwlock_yield_upgrade, &c1_done); in test_co_rwlock_upgrade()
314 qemu_coroutine_enter(c1); in test_co_rwlock_upgrade()
317 /* c1 now should go to sleep. */ in test_co_rwlock_upgrade()
318 qemu_coroutine_enter(c1); in test_co_rwlock_upgrade()
372 * | c1 | c2 | c3 | c4 |
399 Coroutine *c1, *c2, *c3, *c4; in test_co_rwlock_downgrade() local
403 c1 = qemu_coroutine_create(rwlock_rdlock_yield, &c1_done); in test_co_rwlock_downgrade()
408 qemu_coroutine_enter(c1); in test_co_rwlock_downgrade()
413 qemu_coroutine_enter(c1); in test_co_rwlock_downgrade()
419 qemu_coroutine_enter(c1); in test_co_rwlock_downgrade()