A recent video describes a pair of long-awaited developments that promise dramatic benefits in achievable navigation and tracking performance. Marked improvements will occur, not only in accuracy and availability; over four decades this topic has arisen in connection with myriad operations, many documented in material cited from other blogs here.
Let me begin with a quote worth repeating — “Do we really need to wait for a catastrophe before taking action against GNSS vulnerabilities ?” — and follow with an extension of scope beyond.
It’s encouraging to see LinkedIn discussions recognizing ADSB limitations that preclude dependable collision avoidance capability – but that recognition needs to be far more widespread. The limitations are both severe and multifaceted including, in addition to vulnerability from inadequate security,
* accuracy goals based on present position instead of the monumentally more important relative velocity — ADSB allows 10 meter/sec velocity error (!), without characterization as vectorial or relative or probabilistic.
* the glaring but near-universal flaw of sharing coordinates, thereby failing to exploit what made differential operation spectacularly successful: work with individual measurements separately.
Note that these deficiencies existed long before the emergence of unmanned vehicles. The need to correct them is as fundamental as it is urgent. I’ve communicated these concerns over and over, most recently receiving a gratifying response from my June 11 presentation to the satnav National Advisory Board, with details available from URLs at the end.
In that presentation I cited a successful flight validation achieving accuracy on the order of cm/sec, for the crucially important relative velocity between vehicles that can be on or near a collision course. That is a thousand times less error than the 10 meter/sec allowed by ADSB. Furthermore, reduction by a thousand in each of three directions translates into a billion times less volume of uncertainty — or, in just two dimensions at fixed altitude, a million times less area. To realize this crucial safety improvement no new discoveries are needed and no new equipment needs to be invented; only the content of transmitted data needs to change: measurements rather than coordinates. Yet usage of the method is not being planned. After initially proposed before 2000, a limited support program started within the past few years is the only step taken toward this direction.
No claim is made that the last word has been spoken or that introduction of the needed modifications — nor accompanying regulation — would be trivial. The intent here is not criticism and complaints for the sake of criticism and complaints. Emphasizing unwelcome reality always caries risk of drawing wrath. Nevertheless, especially now with growing usage of unmanned vehicles, sounding an alarm is better than passively waiting for a calamity. So here’s an alarm: Inadequate preparation for collision avoidance is a microcosm of a much wider overall flaw in today’s decision-making process. For years substantial numbers of qualified people have spent extensive effort trying to prevent cataclysmic failures in one area or another involving PNT (position/navigation/timing). They definitely deserve attention and action.
Anything approaching a thorough compilation of worthy advocacy would require considerable length; just a few recent examples are cited here. Explanations tracing inaction to current shortcomings can logically include a diagnosis of dissatisfaction expressed at a pinnacle of authority within DoD. An even more current offering is only the latest expression of regret over insufficient support for satnav, describing a highly relevant chain of inaction over a multiyear period. Near the beginning of that period, a cover story for Coordinates magazine repeated a quote from the previous month’s cover story The quote worth repeating, cited at the start of this, is a perfect expression of the frustration prevalent over a decade following the universally acclaimed 2001 Volpe report. Now, almost a decade-and-a-half after that report, partial progress toward a solution coexists with minimal progress toward collision avoidance — while unmanned vehicles are already threatening passenger flight safety. Now to extend the quote: “Do we really need to wait for a catastrophe before making better use of measurements — GNSS or otherwise — to prevent collisions in the presence of increased manned and unmanned traffic?”
Secretary of Defense Ashton Carter’s recent statement “I hate GPS” naturally creates much concern within the navigation community. The July-August issue of InsideGNSS contains his presentation with the reaction from editor Glen Gibbons, plus my own response which delineates
* where the Secretary is badly mistaken, and
* where his concerns are legitimate.
There is a connection between the latter and our industry’s decades-long determined resistance to common-sense improvements in both performance and economy. Steps offering dramatic benefits are further described in material long available from this site. Rather than repeat those descriptions here, I now focus instead on another kind of avoidance: an urgent need to swerve away from another administrative blunder.
Recent history illustrates how the preceding expression is no exaggeration. Loss of LORAN wasn’t permanent, for reasons that were primarily capricious. Planned destruction of vital backup to a vulnerable pillar for communication and navigation wasn’t completed because the government never got around to finishing it. Hundreds of experienced professionals offered testimony in 2009 (including my “two cents’ worth” revisited ) — which failed at the time. Administrative action shut down LORAN for years, with intent to destroy it.
Poor judgment, however, is not the sole cause of unwise administrative action. Often it is prompted by poor performance; the GAO-08-467SP report provides a perfect explanation of that. Dismal as it is, it must be believed that even gross departures from responsible stewardship can be corrected. Destroying a critical resource is obviously not the answer.
Changes in coordinates at stations affected by earthquakes have been monitored successfully, for years, with precision using satellite navigation. Results of interest have then been produced in the past by processing the outcome, e.g., investigating the history of triangles formed thereby. The first application to earthquakes of entirely different criteria (affine deformation states) has produced results with encouraging prospects for prediction, both in time (more than two weeks prior to the 2011 Tohoku quake) and spatially (departures from the affine model at the station nearest epicenter).
The fifteen independent states of a standard 3-D 4×4 affine transformation can be categorized in five sets of three, each set having x-, y-, and z-components for translation, rotation, perspective, scaling, and shear. Immediately the three degrees-of-freedom associated with perspective are irrelevant for purposes here. In addition both translation and rotation, clearly having no effect on shape, can be analyzed separately — and the same is likewise true of uniform scaling. It is thus widely known that there are five “shape states” involved in 3-D affine deformation, three for shear and two for nonuniform scaling. One way to describe shape states is to note their effects in 2-D, where there is only one for nonuniform scaling (which deforms a square into a rectangle) and one for shear (which deforms a rectangle into a parallelogram). Therefore it is noted here that added insight into earthquake investigation can be obtained by analyzing affine features – with specific attention given to their individual traits (degrees-of-freedom).
For investigating earthquakes from affine degrees-of-freedom, methodology of another very different field — anatomy — is highly relevant but ironically lacking a crucial feature. As currently practiced, physiological studies of affine deformations concentrate heavily on two-dimensional representations. While full affine representation is very old, its inversion — i.e., optimal estimation of shape states from a given overdetermined coordinate set — has previously been limited to 2-D.
Immediately then, extension was required for adaptation. The fundamentals, however, still remain applicable. Instead of designated landmark sets coming from a group of patients, here they are associated with a series of days (e.g., from several days before to several days after a quake). Each landmark set is then subjected to a series of procedures (centroiding, normalization, rotation) for fitting landmark sets from one day to another.
The Procrustes representation from procedural steps just described provides sequences of centroid shifts in each direction, rotations about each axis, and amounts of uniform scaling needed for each separate day. In addition to those seven time histories, there are five more that offer potential for greater insight (again, shape states — three shear and two nonuniform scaling). All were obtained for landmark coordinate sets reported before and after the 2011 Tohoku quake. From sample recorded coordinates provided along with shape state values, readers of the manuscript are enabled to verify results.
A video completed recently provides just enough matrix theory needed for Kalman filtering. It’s available for (1) purchase or 72-hour rent at low cost or (2) free to those attending courses I teach in 2014 or after (because the short durations don’t allow time to cover it). The one-hour presentation is divided into three sections. Each section has a preview, freely viewable.
The first section, with almost NO math, begins by explaining why matrices are needed — and then immediately emphasizes that MATH ALONE IS NOT ENOUGH, To drive home that point, a dramatic illustration was chosen. Complex motions of a satellite, though represented in a MATHEMATICALLY correct way, were not fully understood by its designers nor by the first team of analysts contracted to characterize it. From those motions, shown with amplitudes enlarged (e.g., doubled or possibly more) for easy visualization, it becomes clear why insight is every bit as important as the math.
For some viewers the importance of insight alone will be of sufficient interest with no need for the latter two sections. Others, particularly novices aspiring to be designers, will find the math presentation extremely helpful. Straight to the point for each step where matrices are applied, it is just the type of information I was earnestly seeking years ago, whole “pulling teeth” to extract clarification of ONLY NECESSARY theory without OVER simplification.
The presentation supplies matrix theory prerequisites that will assist aspiring designers in formulating linear(ized) estimation algorithms in block (weighted least squares) or sequential (recursive Kalman/EKF) form. Familiar matrix types (e.g., orthogonal, symmetric), their properties, how they are used — and why they are useful — with interpretation of physical examples, enable important operations both powerful and versatile. An enormous variety of applications involving systems of any order can be solved in terms of familiar expressions we saw as teenagers in college.
Useful for either introduction or review, there is no better way to summarize this material than to repeat one word that matters beyond all else — INSIGHT.
The ago-old Interrogation/Response method for air surveillance was aptly summarized in an important 1996 GPSWorld article by Garth van Sickle: Response from an unidentified IFF transponder is useful only to the interrogator that triggered it. That author, who served as Arabian Gulf Battle Force operations officer during Desert Storm, described transponders flooding the air with signals. Hundreds of interrogations per minute in that crowded environment produced a glut of r-f energy – but still no adequate friendly air assessment.
The first step toward solving that problem is a no-brainer: Allocate a brief transmit duration to every participant, each separate from all others. Replace the Interrogation/Response approach with spontaneous transmissions. Immediately, then, one user’s information is no longer everyone else’s interference; quite the opposite: each participant can receive every other participant’s transmissions. In the limit (with no interrogations at all), literally hundreds of participants could be accommodated. Garble nonexistent. Bingo.
Sometimes there a catch to an improvement that dramatic. Fortunately that isn’t true of this one. A successful demo was performed at Logan Airport – using existing transponders with accepted data formatting (extended squitter), in the early 1990s, by Lincoln Labs. I then (first in January 1998) made two presentations, one for military operation (publication #60- click here) and another one for commercial aviation (publication #61-click here), advocating adoption of that method with one important change. Transmitting GPS pseudoranges rather than coordinates would enable an enormous increase in performance. Reasons include cancellation of major errors – which happens when two users subtract scalar measurements from the same satellite, but not coordinates formed from different sets of satellites. That, however, only begins to describe the benefit of using measurements (publication #66); continue below:
With each participant receiving every other participant’s transmissions, each has the ability to track all others. That is easily done because
(1) every extended squitter message includes unique source identification, and (2) multiple trackers maintained in tandem have been feasible for years; hundredsof tracks would not tax today’s computing capability at all. Tracks can be formed by ad hoc stitching together coordinate differences, but accuracy will not be impressive. A Kalman tracker fed by those coordinate differences would not only contain the uncancelled errors just noted, but nonuniform sensitivities, unequal accuracies, and cross-axis correlations among the coordinate pseudomeasurement errors would not be taken into account. Furthermore, the dynamics (velocity and acceleration) – as derivatives – would degrade even more – and dynamic accuracy is absolutely crucial for ability to anticipate near-future position (e.g., for collision avoidance).
The sheer weight of all the considerations just noted should be more than enough to motivate the industry towards preparing to exploit this capability. But, wait – there’s more. Much more, in fact. For how many years have we been talking about consolidating various systems, so that we wouldn’t need so many different ones? Well, here’s a chance to provide both 2-dimensional (runway incursion) and 3-dimensional (in-air) collision avoidance with the same system. The performance benefits alone are substantial but that plan would also overcome a fundamental limitation for each –
* Ground: ASDE won’t be available at smaller airports
* In-air: TCAS doesn’t provide adequate bearing information; conflict resolution is performed with climb/dive.
The latter item doesn’t make passengers happy, especially since that absence of timely and accurate azimuth information prompts some unnecessary “just-in-case” maneuvers.
No criticism is aimed here toward the designers of TCAS; they made use of what was available to them, pre-GPS. Today we have not just GPS but differential GPS. Double differencing, which revolutionized surveying two decades ago, could do the same for this 2-D and 3-D tracking. The only difference would be absence of any requirement for a stationary reference. All positions and velocities are relative – exactly what the doctor ordered for this application.
OK, I promised – not just more but MUCH more. Now consider what happens when there aren’t enough satellites instantaneously available to provide a full position fix meeting all demands (geometry, integrity validation): Partial data that cannot provide instantaneous position to be transmitted is wasted (no place to go). But ancient mariners used partial information centuries ago. If we’re willing to do that ourselves, I’ve shown a rigorously derived but easily used means to validate each separate measurement according to individual circumstances. A specific satellite might give an acceptable measurement to one user but a multipath-degraded measurement to another. At each instant of time, any user could choose to reject some data without being forced to reject it all. My methods are applicable for any frequency from any constellation (GPS, GLONASS, GALILEO, COMPASS, QZSS, … ).
While we’re at it, once we open our minds to sharing and comparing scalar observations, we can go beyond satellite data and include whatever our sensors provide. Since for a half-century we’ve known how to account for all the nonuniform sensitivities, unequal accuracies, and cross-axis correlations previously mentioned, all incoming data common to multiple participants (TOA, DME, etc.) would be welcome.
So we can derive accurate cross-range as well as along-range relative dynamics as well as position, with altitude significantly improved to boot. Many scenarios (those with appreciable crossing geometry) will allow conflict resolution in a horizontal plane via deceleration – well ahead of time rather than requiring a sudden maneuver. GPS and Mode-S require no breakthrough in inventions, and track algorithms already in public domain carry no proprietary claims. Obviously, all this aircraft-to-aircraft tracking (with participants in air or on the ground) can be accomplished without data transmitted from any ground station. All these benefits can be had just by using Mode-S squitter messages with the right content.
There’s still more. Suppose one participant uses a different datum than the others. Admittedly that’s unlikely but, for prevention of a calamity, we need to err on the side of caution; “unlikely” isn’t good enough. With each participant operating in his own world-view, comparing scalar measurements would be safe in any coordinate reference. Comparing vectors with an unknown mismatch in the reference frame, though, would be a prescription for disaster. Finally, in Chapter 9 of GNSS Aided Navigation & Tracking I extend the approach to enable sharing observations of nonparticipants.
In the About panel of this site I pledged to substantiate a claim of dramatic improvements afforded by methods to be presented. This operation is submitted as one example satisfying that claim. Many would agree (and many have agreed) that the combined reasons given for the above plan is compelling. Despite that, there is no commitment by the industry to pursue it. ADSB is moving inexorably in a direction that was set years ago. That’s a reality – but it isn’t the only reality. The world has its own model; it doesn’t depend on how we characterize it. It’s up to us to pattern our plans in conformance to the real world, not the other way around. Given the stakes I feel compelled to advocate moving forward with a pilot program of modest size – call it “Post-Nextgen” – having the robustness to recover from severe adversity. Let’s get prepared.
In 2013 a phone presentation was arranged, for me to talk for an hour with a couple dozen engineers at Raytheon. The original plan was to scrutinize the many facets and ramifications of timing in avionics. The scope expanded about halfway through, to include topics of interest to any participant. I was gratified when others raised issues that have been of major concern to me for years (in some cases, even decades). Receiving a reminder from another professional, that I’m not alone in these concerns, prompts me to reiterate at least some aspects of the ongoing struggle — but this time citing a recent report of flight test verification.
The breadth of the struggle is breathtaking. The About panel of this site offers short summaries, all confirmed by authoritative sources cited therein, describing the impact on each of four areas (satnav + air safety + DoD + workforce preparation). Shortcomings in all four areas are made more severe by continuation of outdated methods, as unnecessary as they are fundamental, Not everyone wants to hear this but it’s self-evident: conformance to custom — using decades-old design concepts (e.g., TCAS) plus procedures (e.g., position reports) and conventions (e.g., interface standards — guarantees outmoded legacy systems. Again, while my writings on this site and elsewhere — advocating a different direction — go back decades, I’m clearly not alone (e.g., recall those authoritative sources just noted). Changing more minds, a few at a time, can eventually lead to correction of shortcomings in operation.
We’re not pondering minor improvements, but dramatic ones. To realize them, don’t communicate with massaged data; put raw data on the interface. Communicate in terms of measurements, not coordinates — that’s how DGPS became stunningly successful. Even while using all the best available protection against interference, (including anti-spoof capability), follow through and maximize your design for robustness; expect occurrences of poor GDOP &/or less than a full set of SVs instantaneously visible. Often that occurrence doesn’t really constitute loss of satnav; when it’s accompanied by history of 1-sec changes in carrier phase, those high-accuracy measurements prevent buildup of position error. With 1-sec carrier phase changes coming in, the dynamics don’t veer toward any one consistent direction; only location veers during position data deficiencies (poor GDOP &/or incomplete fixes) and, even then, only within limits allowed by that continued accurate dynamic updating. Integrity checks also continue throughout.
So then, take into account the crucial importance of precise dynamic information when a full position fix isn’t instantaneously available. Take what’s there and stop discarding it. Redefine requirements to enable what ancient mariners did suboptimally for many centuries — and we’ve done optimally for over a half-century. Covariances combined with monitored residuals can indicate quality in real time. Aircraft separation means maintaining a stipulated relative distance between them, irrespective of their absolute positions and errors in their absolute positions. None of this is either mysterious or proprietary, and none of this imposes demands for huge budgets or scientific breakthroughs — not even corrections from ground stations.
A compelling case arises from cumulative weight of all these considerations. Parts of the industry have begun to address it. Ohio University has done flight testing (mentioned in the opening paragraph here) that validates the concepts just summarized. Other investigations are likely to result from recent testing of ADSB. No claim is intended that all questions have been answered, but — clearly — enough has been raised to warrant a dialogue with those making decisions affecting the long term.
A comment challenged my video . I’m glad it included an acknowledgment that some points might have been missed. To be frank that happened a bunch; bear with me while I explain. First, there’s the accuracy issue; doppler &/or deltarange info provided from many receivers is far less accurate than carrier phase (sometimes due to cutting corners in implementation — recall that carrier phase, as the integral of doppler, will be smoother if processing is done carefully). Next, preference for 20-msec intervals will backfire badly. If phase noise at L-band gives a respectable 7mm = 0.7cm, doppler velocity error [(current phase) – (previous phase)] / 1 sec is (1.414) (0.7) = 1 cm/sec RMS for a 1-sec sequential differencing interval. Now use 20 msec: FIFTY times as much doppler error! Alternatively if division is implicit instead of overt, degradation is more complicated: sequential phase differences are highly correlated (with a correlation coefficient of -1/2, to be precise). That’s because the difference (current phase) – (previous phase) and the difference (next phase) – (current phase) both contain the common value of current phase. In a modern estimation algorithm, observations with sequentially correlated errors are far more difficult to process optimally. That topic is a very deep one; Section 5.6 and Addendum 5.B of my 2007 book address it thoroughly. I’m not expecting everyone to go through all that but, to offer fortification for its credibility, let me cite a few items:
* agreement from other designers who abandoned efforts to use short intervals
* table near the bottom of a page on this site.
* phase residual plots from Chapter 8 of my 2007 book.
The latter two, it is recalled, came from flight test for an extended duration (until flight recorder was full), under severe test aircraft (DC-3) vibration.
For doppler updating from sources other than satnav, my point is stronger still. Doppler from radar (which lacks the advantage of passive operation) won’t get velocity error much below a meter/sec — and even that is an improvement over unaided inertial nav (we won’t see INS velocity specs expressed in cm/sec within our lifetime).
Additional advantages of what the video offers include (1) no requirement for a mask angle (2) GNSS interoperability, and (3) robustness. A brief explanation:
(1) Virtually the whole world discards all measurements from low-elevation satellites because of propagation errors. But ionospheric and tropospheric effects change very little over a second; 1-sec phase differences are great for velocity information. Furthermore they offer a major geometry advantage while occurrence of multipath would stick out like a sore thumb, easily edited out.
(2) 1-sec differences from various constellations are much easier to mix than the phases themselves.
(3) For receivers exploiting FFT capability even short fragments of data, not sufficiently continuous for conventional mechanizations (track loops), are made available for discrete updates.
The whole “big picture” is a major improvement is robust operation
The challenger isn’t the only one who missed these points; much of our industry, in fact, is missing the boat in crucial areas. Again I understand skepticism, but consider the “conventional wisdom” regarding ADSB: Velocity errors expressed in meters per second — you can hear speculative values as high as ten. GRADE SCHOOL ARITHMETIC shows how scary that is; collision avoidance extrapolates ahead. Consider the vast error volume resulting from doing that 90 seconds ahead of closest approach time with several meters per second of velocity error. So — rely on see-and-avoid? There are beaucoup videos that show how futile that is (and many more videos that show how often near misses occur — in addition there are about a thousand runway incursions each year). That justifies the effort for dramatic reduction of errors in tracking dynamics — to cm/sec relative velocity accuracy.
It’s perfectly logical for people to question my claims if they seem too good to be true. All I ask is follow through, with visits to URLs cited here.
A LinkedIn discussion centered on the Future Airborne Capability Environment (FACE) standard contained an important observation concerning certification. Granted — requirements for validation, with acceptance by governing agencies, definitely are essential for safety. What follows here is advocacy for a proposed way to realize the common avionics benefits offered by FACE while retaining (and in fact, improving) the process of certification. Reasoning is based on three major items:
* CHANGE. In many respects this has necessitated improved standards.
* HISTORY. Spectacular failures in what we have now are widely documented.
* COST. The status quo is (and, for a long time, has been) unaffordable.
In regard to the first item: the pace of change in so many areas (hardware, software, operating systems, data communication, etc., etc., etc.) — and the effects on procurement cycles — are well known. How can certification remain unchanged when nothing else does? That argument would be undercut if the process had a rock solid track record — but that theme would not be supported by the second item — history:
Myriad shortcomings of existing operational systems are so pervasive that no one is considered a “loose cannon” for openly discussing them. Any of my horror stories — too strange and too numerous to be revisited here — would be trumped anyway by a document from the government itself. GAO-08-467SP, in 2008, described outlandish cost overruns, schedule delays, and deficient technical performance in the defense industry. That 3-way combination speaks for itself. Now a significant addition: the certification process has not been at all immune to serious flaws. The first-ever certified GPS receiver is now well known to have failed spectacularly in multiple facets of integrity testing by another manufacturer. It is readily acknowledged that correction of those early problems is quite credible, but one issue is inescapable: Historical proof of flightworthiness improperly bestowed — with proprietary rights accepted for algorithms and tests –- happened, and that was not widely known until much later.
There is still more, including integrity failure probability limits missed by orders-of-magnitude in certified GPS receivers, severe limitations of GO/NO-GO testing, and failed attempts to gain approval to set requirements for correcting those plus other deficiencies. For brevity here, those issues are covered by citing the fifth page from another related reference.
The final item is, after years of fruitless talk about cost reduction, being acknowledged — we can’t do what we’ve been doing any more. With dollars being the ultimate driver of so many decisions, we might finally see the necessary break from ingrained habits. FACE already addresses the issues and the requisite justifications. To make it all happen, two essential ingredients are
* raw-data-across-the-board, and
* nonproprietary software, with standardization under government control.
Flight-validated algorithms already in existence can be converted (e.g., from proof-of-concept to in-flight real-time form) according to government specification, by small groups more interested in engineering than in dollars (yes, that does exist). The payoff in cost savings can be huge.
Significant momentum is evolving toward a role for Open System Architecture (OSA) applied to radar. My observations in connection with that, voiced in a short LinkedIn discussion, seem worth repeating here.
One step could add major impact to this development: Rather than position (or relative position) outputs, provide measured range, azimuth, elevation (doppler could optionally be added if applicable) on the output interface. That simple step would vastly improve effectiveness of track file maintenance. Before attempting to describe all reasons for improved performance, two obvious benefits can be identified first:
* ability to use partial information (e.g., range-only or, for passive operation, angle-only)
* proper weighting of data for updating track state estimates.
The first item is self-evident. The second arises from common-sense attachment of greater value to the most accurate information. An explanation:
One-sigma error ellipsoids for individual radar fixes are not spherical (not a beachball shape but more like a flattened beachball), even at close range. At longer distances the shape progresses from a frisbee to a pancake to a DVD. Kalman filtering has enabled us to capitalize on that feature for over a half-century. Without exploiting it, we effectively treat separate radar-derived “coordinates” by intersecting volumes in space that are common to overlapping spheres. Resulting uncertainty volume is enormously larger than it should be.
The feature just noted shows up dramatically when mixing data among multiple platforms. Consider cooperative engagement whereby participants, all tracking each other via network-transmitted GPS observations, share radar observations from an unknown non-participant. Share measurements or coordinates? No contest; multiple lines crossing from different directions can offer best (i.e., along-range) accuracies applicable in 3-D.
That fact (i.e., combining data from different sensors and different platforms further dramatizes available improvements) doesn’t diminish the basic issue; even with a time history of data from only one radar, a designer with direct measurements available — instead of, not in addition to, coordinates — can provide incomparably superior performance.
“Send Measurements not Coordinates” (1999; #66 from the “Published Articles” panel, opening with eight rock-solid reasons) was aimed at GPS rather than radar. Many of the principles are the same when mixing data with information from other platforms — and from other sensors such as GPS. There is no reason, in fact, why data from satellite navigation and radar can’t be combined in the same estimation algorithm. That practice hasn’t evolved but the historical separation of operations (e.g., navigation and surveillance), arising from old equipment limitations, should no longer be a constraint. Moreover, with focus shifted from tracking to navigation, integration with additional (e.g., inertial) data offers still more reasons for using direct measurements. Rather than loose integration, superior benefits are widely known to result as the sophistication progresses forward (tight. ultratight, and deep integration).
Further elaborations on “casting off our old habits” appear from different perspectives in various blogs, one-pagers, and a few manuscripts available at this site. If your library has a copy of GNSS Aided Navigation & Tracking pages 203-4 show a way to implement the cooperative sharing of radar data obtained from a non-participant, among participants tracking each other via the mutual surveillance and tracking approach defined earlier in that same chapter.
Because so many operational systems (in fact, a vast majority) use reports in the form of coordinates, reiteration is warranted. The central issue is the content, not the amount, of data. Rather than coordinates, provide accurately time-stamped direct measurements with links connected to whichever platform observed the data (e.g., for satnav — pseudoranges; for radar — range, azimuth, elevation). Those links are automatically attached when Mode-S extended squiter (e.g., chosen for ADSB) is the means for conveying the data. For message content, strictly disallow “massaging the data beyond the light of day” (e.g., by unknown processes, with uncertain timing, … ) which invariably results in enormous loss of performance in common occurrence today.