What is the number of one second ticks between Unix time epoch (01 Jan 1970) and GPS time epoch (06 Jan 1980)?
I have seen multiple answers from several sources on t
What is the number of one second ticks between Unix time epoch (01 Jan 1970) and GPS time epoch (06 Jan 1980)?
There are at least two possible answers:
What is POSIX timestamp for 1980-01-06 UTC? Answer: 315964800 (exactly), in Python:
from datetime import datetime, timedelta
print((datetime(1980,1,6) - datetime(1970,1,1)) // timedelta(seconds=1))
It is the number of SI seconds between the dates not counting leap seconds. In other words, the code shows how many "UT1" seconds (~1/86400 of a mean-solar day) passed between the events.
UTC, GPS time scales tick in SI seconds. UTC "forgets" leap seconds and therefore the actual number of SI seconds between the dates is slightly larger than the POSIX timestamp.
315964800
is not the correct answer if you want to find elapsed seconds
How many SI seconds elapsed between 1970-01-01 UTC and 1980-01-06 UTC? Answer: 315964811 (approximately).
To answer the second question, you need to know how many intercalary leap seconds were inserted between the two dates (convert UTC to the International Atomic Time (TAI)):
#!/usr/bin/env python3
from datetime import datetime, timedelta
tai_posix_epoch = datetime(1970, 1, 1) + timedelta(seconds=8, microseconds=82)
tai_gps_epoch = datetime(1980, 1, 6) + timedelta(seconds=19)
print(round((tai_gps_epoch - tai_posix_epoch) / timedelta(seconds=1)))
The difference between TAI and GPS time is constant within 10s of nanoseconds.
The time between 1970 and 1972 (when UTC was introduced) is a little fuzzy; the TAI-UTC difference is not integer number of seconds in that period:
from decimal import Decimal as D
MJD_1970_01_01 = 40587
dAT_1970_01_01 = D("4.213170") + (MJD_1970_01_01 - 39126) * D("0.002592")
# -> delta(AT) = TAI - UTC = Decimal('8.000082') # 8 seconds, 82 microseconds
Here's a picture that shows the relation between UT1, UTC, and TAI time scales over the years: Each step is a leap second starting with TAI - UTC = 10s on 1972-01-01. 26 positive leap seconds had been inserted as of 1 July 2015.
315964819
timestamp could be explained if 1970-01-01 00:00:00 TAI
epoch is used:
print(datetime(1970, 1, 1) + timedelta(seconds=315964819)) # TAI
# 1980-01-06 00:00:19 TAI or 1980-01-06 00:00:00 UTC
i.e., exactly 315964819
SI seconds elapsed between 1970-01-01 00:00:00 TAI
and 1980-01-06 00:00:00 UTC
(note: the dates are expressed using different time scales).
"right" timezones use 1970-01-01 00:00:10 TAI
epoch (notice: 10 seconds) and therefore the corresponding timestamp for the GPS epoch (1980-01-06 00:00:00 UTC) is 315964809
(not 315964819
). Here's a succinct description of the difference between "right" and POSIX timestamps:
The "right" files in the tz (zoneinfo) database have a subtle difference from the POSIX standard. POSIX requires that the system clock value of
time_t
represent the number of non-leap seconds since 1970-01-01. This is the same as requiring POSIX seconds to be mean solar seconds of UT, not the atomic seconds that UTC has counted since 1972-01-01.The "right" zoneinfo files assert that the system clock value of
time_t
represent the actual number of seconds in the internationally approved broadcast time scale since 1970-01-01. As a result the value of time_t which is expected by the "right" zoneinfo files is greater than the value of time_t specified by POSIX. The difference in the values of time_t is the number of leap seconds which have been inserted into the internationally approved broadcast time scale. As of year 2015 the difference is 26 seconds.emphasize is mine
Can someone authoritative please step in here?
IERS BULLETIN C (the data that I've used above) is the authority on leap seconds (and therefore (indirectly) on the difference between UTC and GPS time scales).
I guess I'm in the third camp :)
Let's call it like it is:
2,904,548,141,415,381,930 "periods[...]of a caesium 133 atom" measured at 0 degrees Kelvin at the geoid. (give or take a few hundred million periods depending on which TAI/SI definition you use)
It depends on what time scales (and which definitions of those time scales) you're using.
315964809 in TAI seconds (1977 definition) and thus UTC seconds
315964800 in UNIX seconds
(both are equal to eachother but ONLY between your specified dates and both correspond to 2,904,548,141,415,381,930 "periods[...]")
Please note that UNIX seconds replay the same second after the completion of a UTC leap second, so that the UTC seconds, 2012-06-30 23:59:60 UTC and 2012-07-01 00:00:00 UTC, were both represented by a UNIX timestamp of 1341100800.
Using TAI seconds
Even though they aren't really, let's assume that all TAI seconds before 1977 are still exactly equal to the 1977/1997 definition of TAI/SI seconds.
Let's also assume that by
"Unix time epoch (01 Jan 1970)" to "GPS time epoch (06 Jan 1980)"
you mean
1970-01-01 00:00:10 TAI to 1980-01-06 00:00:19 TAI
in this case there would be
( ( (365days/year * 10years) + 2 leap days + 5 days) * 86400 TAI seconds/day ) + 9 TAI seconds
= 315964809 TAI seconds
Using UNIX seconds
Even though they aren't really, let's assume that the duration of a UTC second before 1977 is still exactly equal to the 1977/1997 definition of a TAI/SI second.
Let's also assume that by
"Unix time epoch (01 Jan 1970)" to "GPS time epoch (06 Jan 1980)"
you mean
1970-01-01 00:00:00 UTC to 1980-01-06 00:00:00 UTC
and that UNIX time skips back a second after the completion of a leap second
in this case there would be
( ( (365days/year * 10years) + 2 leap days + 5 days) * 86400 seconds/day ) + 9 leap seconds - 9 unix leap second rehashes
= 315964800 UNIX seconds
Concerning "Periods[...]"
A 1977/1997 TAI/SI second is what was used to come up with 315964809 seconds of 9,192,631,770 periods each = 2,904,548,141,415,381,930 periods. An 1997 SI second is equal to the duration of 9,192,631,770 periods of the radiation corresponding to the transition between the two hyperfine levels of the ground state of the caesium-133 atom at rest at a temperature of 0 K. The 1977 definition of TAI measures SI seconds at the geoid.
The different values you stated are caused by mixing up the 1970 to 1980 offset with leap seconds.
The correct offset value is 315964800 seconds.
Explanation:
UTC and GPS time deviate (on average) every 18 months by one additional second.
This is called a leap second, introduced in UTC time base, necessary to adjust for changes in the earth's rotation.
GPS Time not adjusted by leap seconds.
Currently (2013) there is an offset of 16s:
GPS Time-UTC = 16 seconds
Unix time is a time format not a time reference. It represents the number of milliseconds (or seconds) since 1.1.1970 UTC. Ideally your system time is synchronized with UTC by a TimeServer (NTP).
To convert, and get your offset, you should use a fixed offset: (6.1.1980 UTC - 1.1.1970 UTC)
and THEN add the current value of GPS to UTC deviation (currently 16s). E.g make that value configurable, or read the current offset from a GPS device (they know the difference between UTC and GPS Time)
The different values you stated are caused by mixing up 1970 to 1980 offset with leap seconds. Dont do that, handle them separately.
This java program:
SimpleDateFormat df = new SimpleDateFormat();
df.setTimeZone(TimeZone.getTimeZone("UTC"));
Date x = df.parse("1.1.1970 00:00:00");
Date y = df.parse("6.1.1980 00:00:00");
long diff = y.getTime() - x.getTime();
long diffSec = diff / 1000;
System.out.println("diffSec= " + diffSec);
Outputs this value:
diffSec= 315964800
So this is the correct offset between 1.1.1970 UTC and 6.1.1980 UTC where GPS Time began. Then you have to correct further 16 seconds which were introduced since 6.1.1980 and today, to calculate the GPS Time of a current UTC time.
go back to the orginal question, "Unix time epoch (01 Jan 1970)" to "GPS time epoch (06 Jan 1980)" , so the epoch is 315964800, 315964819 is the 'TAI epoch' to the 'GPS time epoch'. which means 315964819= 315964800 + 19. so the epoch value you use in the code, is really depending on which time epoch you are using.