|Doc. No.:||WG21/P0250R3 (Kona revision)|
|Author:||Hans-J. Boehm, further edits by Jens Maurer|
This is an attempt to turn the Kona SG1 discussion of CWG 2046 into wording. This revision also reflects SG1 discussion in Jacksonville, Oulu, and Issaquah, along with an initial CWG review in Issaquah.
In several places the standard needs to require that some operation, e.g. an initialization, happens before another one, in roughly, but not quite, the sense of the 1.10.1 memory model. The difference is that we expect this happens before relation to compose with other happens-before relationships. But our normal happens before relation no longer composes in this way, due to the presence of memory_order_consume. Hence we suggest introducing the required notion in 1.10.1.
There was some discussion in Jacksonville about using "synchronizes with" instead of introducing new terminology. That appears to be technically workable, but rather misleading. "Synchronizes with" is currently only defined between actions by different threads, and does not include sequencing. Many of the guarantees we would like to provide will often rely on sequencing rather than synchronization. Thus my current feeling is to introduce new terminology.
The current standard sometimes uses "sequenced before" to express this kind of relationship. This is incorrect and repaired below, since sequencing is a relation on evaluations within a single thread.
In Kona 2015, SG1 rather quickly concluded that a C++ program should be allowed to construct multiple threads for constructors etc., even if the program itself did not use threads. That is not reflected here, mostly because it is a more profound issue than we appreciated at the time.
If we went this route, we would break compatibility with C++03, in what may be a rather significant way. Static constructors for old single-threaded programs would be required to be thread-safe. This is unlikely to be automatically true. It is not that uncommon for constructors to e.g. add to a global data structure. The order in which items are added may not matter. But such a data structure implemented in a single-threaded program is unlikely to be thread-safe.
In Jacksonville, it became clear that some of the confusion here stems from uncertainty about the motivation for the current wording. Currently, we very explicitly allow static constructors to run after the start of main, whether or not other threads are started. This appears to be motivated by the intent to support e.g. lazily loading a dynamic library when a function symbols is referenced, as with RTLD_LAZY on Posix systems. Even if static namespace-scope constructors are run immediately in library loading, the library may be implicitly loaded after the start of main. This seems inherently unsafe; if a constructor requiring a mutex M is run as part of lazy library loading at a point at which M is held by the loading thread, we unavoidably deadlock. And there seems to be no easy way to avoid that.
The solutions appear to be:
Here we follow the last path, but choose not to be specific, and instead make the precise points at which construction occurs implementation-defined. This seems to better reflect the actual status quo. It also means that users who avoid dynamically initialized static objects can continue to take advantage of facilities like RTLD_LAZY.
Running a constructor while main is running inherently implies some notion of concurrency, at least between constructors and main-line code. In the multi-threaded case, constructors run fully concurrently with whatever non-initializing threads happen to be running at the time. But it really only results in constructors running in existing threads, not the starting of additional threads solely to increase constructor parallelism. This seems to be a bit easier to specify than the original version, which allowed parallel constructor execution once a user thread had been started. SG1 generally feels that static namespace-scope constructors should be avoided, and thus speeding them up is a low priority; thus we decided to restrict such constructors to existing threads, which appears to be consistent with known implementations.
We concluded in Kona 2015 that main did not necessarily have a thread id, and that some implementations went out of their way to keep it from having one. After looking at the N4567 wording again, my conclusion is that this interpretation of the standard is a stretch. Some of this wording may have improved since the implementers looked at it.
The first sentence of 1.10 [intro.multithread] states
A thread of execution (also known as a thread) is a single flow of control within a program, including the initial invocation of a specific top-level function, and recursively including every function invocation subsequently executed by the thread.
188.8.131.52 [thread.thread.id] states
An object of type thread::id provides a unique identifier for each thread of execution and a single distinct value for all thread objects that do not represent a thread of execution (30.3.1).
30.3.2 [thread.thread.this] states
thread::id this_thread::get_id() noexcept;
Returns: An object of type thread::id that uniquely identifies the current thread of execution. No other thread of execution shall have this id and this thread of execution shall always have this id. The object returned shall not compare equal to a default constructed thread::id.
There were some voices in Jacksonville in favor of preparing the standard for execution agents that do not support a thread id (or thread local storage). This would simplify GPU execution, but it seems to require that the standard library specify which calls need thread local storage, and which do not, something we don't currently specify. It is unclear to me that we should address that issue in this context.
Change in 3.6.1 [basic.start.main] paragraph 1:
A program shall contain a global function calledMove 1.9 [intro.execution] paragraph 6 to the end of section 1.9:
, which is the designated start of the program.
Add after 1.10.1p10 [intro.races]:
If a signal handler is executed as a result of a call to the std::raise function, then the execution of the handler is sequenced after the invocation of the std::raise function and before its return. [ Note: When a signal is received for another reason, the execution of the signal handler is usually unsequenced with respect to the rest of the program. — end note ]
... The sequencing constraints on the execution of the called function (as described above) are features of the function calls as evaluated, whatever the syntax of the expression that calls the function might be.
Change the last part of 3.6.2p2 [basic.start.static] as follows:
If constant initialization is not performed, a variable with static storage duration (3.7.1) or thread storage duration (3.7.2) is zero-initialized (8.6). Together, zero-initialization and constant initialization are called static initialization; all other initialization is dynamic initialization. Static initialization
shall be performedbefore any dynamic initialization takes place. [ Note: The dynamic initialization of non-local variables is described in 3.6.3; that of local static variables is described in 6.7. -- end note ]
Add a new paragraph after 3.6.3 [basic.start.dynamic] paragraph 1:
Dynamic initialization of a non-local variable with ... [ Note: ... ]Change the middle two bullet points in 3.6.3p2 [basic.start.dynamic] as follows:
- If V has partially-ordered initialization, W does not have unordered initialization, and V is defined before W in every translation unit in which W is defined, the initialization of V is sequenced before the initialization of W if the program does not start a thread (1.10) and otherwise happens before the initialization of W.
- Otherwise, if
aprogram starts a thread before either V or W is initialized, the initializations of V and W are unsequenced .
For the following, Richard points out that we could possibly combine the wording of the deferred and non-deferred cases. My personal preference would be to keep the phrasing so that it is easy for non-deferred implementations to just say that.
I don't know of a good programming model to deal with deferred initialization of statics in the presence of threads. There are serious deadlock issues brought about by initialization when the initializing thread is holding a lock, and I don't know how to resolve them. I would support prohibiting that altogether, but I'm told a few implementations currently do so.
Change the beginning of 3.6.3p3 [basic.start.dynamic] as follows:
It is implementation-defined whether the dynamic initialization of a non-local non-inline variable with static storage duration
happens beforethe first statement of main If the initialization is deferred to happen after the first statement of main, it happens before the first odr-use (3.2) of any non-inline function or non-inline variable defined in the same translation unit as the variable to be initialized.
Change 3.6.3p4 as follows:
It is implementation-defined whether the dynamic initialization of a non-local inline variable with static storage duration
happens beforethe first statement of main . If the initialization is deferred to happen after the first statement of main, it happens before the first odr-use (3.2) of that variable.
Change 3.6.3p5 as follows to fix what looks like an unrelated bug, and pin down initialization of thread-locals:
Change 3.6.4p1 [basic.start.term] as follows:
It is implementation-defined whether the dynamic initialization of a non-local non-inline variable with
static orthread storage duration is sequenced before the first statement of the initial function of thethread . If the initialization is deferred to some point in time sequenced after the first statement of the initial function of the thread, itis sequenced before the first odr-use of any variable with thread storage duration defined in the same translation unit as the variable to be initialized.
Destructors (12.4) for initialized objects (that is, objects whose lifetime (3.8) has begun) with static storage duration, and functions registered with
atexit, are called
as a result of returning fromas part of the call to
Destructors for initialized objects with thread storage duration within a given thread are called as a result of returning from the initial function of that thread and as a result of that thread calling std::exit. The completions of the destructors for all initialized objects with thread storage duration within that thread
are sequencedbefore the initiation of the destructors of any object with static storage duration. If the completion of the constructor or dynamic initialization of an object with thread storage duration is sequenced before that of another, the completion of the destructor of the second is sequenced before the initiation of the destructor of the first.
If the completion of the constructor or dynamic initialization of an object with static storage duration
is sequencedbefore that of another, the completion of the destructor of the second is sequenced before the initiation of the destructor of the first. [ Note: This definition permits concurrent destruction. -- end note ]If an object is initialized statically, the object is destroyed in the same order as if the object was dynamically initialized. For an object of array or class type, all subobjects of that object are destroyed before any block-scope object with static storage duration initialized during the construction of the subobjects is destroyed. If the destruction of an object with static or thread storage duration exits via an exception, std::terminate is called (15.5.1).
Change 3.6.4p3 [basic.start.term] as follows:
If the completion of the initialization of an object with static storage duration
is sequencedbefore a call to
std::atexit(see <cstdlib>, 18.5), the call to the function passed to
std::atexitis sequenced before the call to the destructor for the object. If a call to
is sequencedbefore the completion of the initialization of an object with static storage duration, the call to the destructor for the object is sequenced before the call to the function passed to std::atexit. If a call to
is sequencedbefore another call to
std::atexit, the call to the function passed to the second
std::atexitcall is sequenced before the call to the function passed to the first
Change nothing in 6.7p4 [stmt.dcl]:
Dynamic initialization of a block-scope variable with static storage duration (3.7.1) or thread storage duration (3.7.2) is performed the first time control passes through its declaration; such a variable is considered initialized upon the completion of its initialization. If the initialization exits by throwing an exception, the initialization is not complete, so it will be tried again the next time control enters the declaration. If control enters the declaration concurrently while the variable is being initialized, the concurrent execution shall wait for completion of the initialization. If control re-enters the declaration recursively while the variable is being initialized, the behavior is undefined. [ Example: ... ]
Change the note in 18.5p13 as follows:
call a registered function froma different thread than the one that registered it, so registered functions should not rely on the identity of objects with thread storage duration. — end note ]