A decoded time is an ordered series of nine values that, taken together, represent a point in calendar time (ignoring leap seconds):
An integer between 0 and~59, inclusive.
An integer between 0 and~59, inclusive.
An integer between 0 and~23, inclusive.
An integer between 1 and~31, inclusive (the upper limit actually depends on the month and year, of course).
An integer between 1 and 12, inclusive; 1~means January, 2~means February, and so on; 12~means December.
An integer indicating the year A.D. However, if this integer is between 0 and 99, the "obvious" year is used; more precisely, that year is assumed that is equal to the integer modulo 100 and within fifty years of the current year (inclusive backwards and exclusive forwards). Thus, in the year 1978, year 28 is 1928 but year 27 is 2027. (Functions that return time in this format always return a full year number.)
An integer between~0 and~6, inclusive; 0~means Monday, 1~means Tuesday, and so on; 6~means Sunday.
A generalized boolean that, if true, indicates that daylight saving time is in effect.
A time zone.
Figure 25–5 shows defined names relating to decoded time.
decode-universal-time get-decoded-time