Professional Documents
Culture Documents
1
Space Robotics Corporation
10860 East Oakwood Drive
Tucson, AZ 85749, USA
bgajjar@space-robotics.com
2
Computer Science
Florida Gulf Coast University
Fort Myers, FL 33928, USA
zalewski@fgcu.edu
1
some hover type UAV might have manipulators, pod. During landing the following steps are
which move in Local spaces, the overall followed:
movement of a robotic UAV can be classified into 1. Arresting wires are stretched on the deck
a Global space. Global movement involves travel manually.
along distances that exceed the total length of the 2. Carrier Air Traffic Controls monitors the fuel
UAV itself. The overall global motion of a UAV level in each of the approaching planes to
on an aircraft carrier can be broken down into an select the ones for immediate landing.
on deck 2D maneuver through cluttered 3. The plane hooks its tail hook in the arresting
environment consisting of personnel, planes, and wires at right angle and lands abruptly
miscellaneous on board hardware like hatches, following landing signals.
opening, ducts, piping, electrical housing, motors
etc. The landing process is guided and monitored by
LSO (Landing Signal Officers). LSO’s help guide
Besides the 2D taxi before the take off and after the plane in, through radio communication as well
landing operation to the service hangar the UAV as a collection of lights on the deck. If the plane is
will have to do a 3D approach right before landing coming in okay, the LSO’s will illuminate green
where it needs to precisely sense the landing path lights to tell the pilot everything is okay. If the
and the obstacles and also take off smoothly. plane is off course, the LSO’s can illuminate other
Small UAV’s with vertical take off and landing lights to correct him or "wave him off" (send him
don’t pose a great deal of problem in around for another attempt). In addition to the
transportation on deck and landing and take off. LSO’s, pilots look to the Fresnel Lens Optical
However for larger UAV’s closer to the size of Landing System, commonly referred to as the lens,
real planes this poses a challenge, which must be for landing guidance (Fig. 2). The lens consists of
handled suitably by some novel sensor technology. a series of lights and Fresnel lenses mounted to a
The Global 3D motion of the UAV includes take gyroscopically stabilized platform. The lenses
off, landing, hovering and cruise mode. focus the light into narrow beams that are directed
into the sky at various angles.
During take off involving manual control the Figure 2. Fresnel landing signals.
following steps are followed:
1. Flight crew deck moves planes in position at
the rear of the catapult and attaches the tow 2. PROBLEM DESCRIPTION
bar, hold bar and the JBD (Jet Blast Deflector)
on the nose gear. Since UAV has no human pilot, the process of
2. Catapult officer (shooter) gets the catapults detecting light signals to land along with on board
ready from the control pod. obstacle navigation on the deck of the aircraft
3. Charged catapults are launched to get the carrier must be automated. The overall objective
aircraft off the deck. is to develop complete sensory suite to allow fully
For a UAV with size approximately similar to a autonomous landing, take-off and on-deck
larger plane the process of catapulting it to provide maneuvering of a UAV with smooth technology
lift in short space will be required hence it must transition.
autonomously guide itself to the catapult launching
2
The specific long-term research objectives are Most of the on-board available hardware on the
enumerated as follows: aircraft carrier, to send out angular information to
1. Investigate the feasibility of developing a an approaching UAV, will be utilized. The UAV
COTS sensor system capable of autonomous control system will then translate these signals
landing and take off of Unmanned Aerial (commonly optical) to navigational commands
Vehicle from a CV. using our strap-on-sensor package and control its
2. Interface a commercial UAV guidance, attitude for a smooth descent.
navigation and control system to SRC strap-
on sensor package to control Carrier Approach The ACLS (Automatic Carrier Landing System)
Variables (CAVs). has been a feature of aircraft-carrier landing
3. Develop sensor system such that minimum systems since 1960’s and uses the shipboard based
changes on the CV deck are required for AN/SPN-46 ACLS precision tracking radar, a
automated landing and take off. general purpose computer (GPC) and a data link
4. Utilize commercial and off-the-shelf transmitter. The “most” automated mode of ACLS
components to develop our sensor system to is the only one, which comes close to our
automatically capture existing signals from approach, in which radar tracking of a beacon
LSO’s and landing lights, minimizing the system in the aircraft establishes aircraft position.
intervention of human LSO’s. Computer software, containing necessary control
5. Develop robust and efficient method for UAV logic generates aircraft vertical rate (pitch) and
to CV identification, attitude estimation and bank. Commands are sent to the aircraft via UHF
recursive sensor calibration along with link. In our system this information (existing) will
automatic feature extraction and remote be used as a reference signal to compare UAV’s
triangulation. imaging sensor based control and actuator turn rate
6. Off-line simulation of the process in Phase I for satisfactory guidance and centerline judgment
developing non-holonomic constraints based to match drop light, Flight Deck Indicators and
model in closed form employing polynomial Meatball lighting.
inputs.
7. Develop algorithms using cooperative
imaging technology, where modulating
actively controlled beacons, landing signals
are captured in real time using CCD’s with a
rise time of few microseconds and compared
as error signals with GPS and radar data.
The pilot will see different lights depending on the Roll-Pitch-Yaw based kinematics used to model an
plane's angle of approach (Fig. 3). If the plane is approaching UAV (landing/takeoff mode) and
right on target, the pilot will see an amber light, planar mobile car-like kinematics are used for
dubbed the "meatball", in line with a row of green modeling UAV movements on the deck before
lights. If the amber light appears above the green takeoff and after landing. For approach, we
lights, the plane is coming in too high; if the amber adopted the global motion (translational and
light appears below the green lights, the plane is rotational) of a single UAV (Dogan and Kabamba,
coming in too low. If the plane is coming in way 2000; Venkataramanan and Dogan, 2003). The
too low, the pilot will see red lights. generalized solution for a single UAV can be
extended to multiple UAV’s. The Center of Mass
We plan to develop a unique sensor package to (CM) of a UAV is defined with respect to the
strap on existing UAV systems and provide 2D inertial reference frame (Fig. 4) affixed to the hull
obstacle avoidance and 3D signal capture and of the aircraft carrier (ignoring small motions of
implementation of landing and take-off algorithms. the boat) and after landing the kinematics are
3
switched to car like configuration with a non- correct approach. Steady approach conditions are
holonomic constraint of pure rolling without given by X 0 , Y0 , Z 0 , 0 , 0 , 0 , about which
slipping.
small perturbations x, y , z and , , are
defined. The corresponding perturbation velocities
in the wind frame are u , v, w and p, q, r , about
steady flight values U 0 ,V0 ,W0 , P0 , Q0 , R0 .
During landing or take off, a linear stabilization is
required with zero roll and yaw velocities
0 ). Furthermore
(
V0 0 0 0
for a symmetric flight. Translational and rotational
dynamics are given by Newton-Euler equations,
for mass m and moments of inertia I and for
Figure 4. Reference frames.
trimmed flight state P0 Q0 R0 0 (McLean
The translation rate of change is rB and 1990; Blakelock 1991).
A simulation of the dynamics of the UAV model Blakelock, J. H. (1991). Automatic Control of
was written using MATLAB. This simulation Aircraft and Missiles, John Wiley and Sons.
models the velocity, altitude, flight path angle, Dogan, A., P.T. Kabamba (2000). “Escaping a
range, pitch, and angle of attack of the UAV. microburst with turbulence: Altitude, dive and
Preliminary work has been completed on a flight pitch guidance strategies”, J. Aircraft, Vol. 37,
path angle autopilot. The autopilot controls the No. 3, May-June 2000, pp. 417-26
flight path angle by using the angle of the elevator McLean, D. (1990). Automatic Flight Control
and the magnitude of thrust. The transfer function Systems, Prentice Hall.
Pachter, M., C.H. Houpis (1996). Flight Control of
( s) Piloted Aircraft, In: The Control Handbook,
of was computed (Fig. 5), and a PID
e( s ) W. Levine (Ed.), CRC Press, Chapter 75.1,
controller was then used to stabilize the system. pp. 1287-1303.
Sasiadek J., Q. Wang, R. Johnson, L. Sun, J.
The response time to a 1.0 deg step input of the Zalewski (2000). UAV Navigation Based on
elevator is shown in Figure 6. It takes about 4 Parallel Extended Kalman Filter, Proc. AIAA
seconds for the system to stabilize. Guidance, Navigation and Control
Conference, Denver, Colorado, August 14-17,
1.6 2000, Paper AIAA-2000-4165
1.4 US DOD (2000) Unmanned Aerial Vehicle (UAV)
Roadmap, Office of the Secretary of Defense,
Flight Path Angle (deg)
1.2
http://www.acq.osd.mil/usd/uav_roadmap.pdf
1
US JFC (2002). Swarming Entities - The
0.8 Operational Utilities of Establishing Humans-
0.6 on-the-Loop, Rapid Assessment Process
Report #02-04, US Joint Forces Command, 3
0.4
June 2002
0.2
Venkataramanan S., A. Dogan (2003). “Nonlinear
0
0 1 2 3 4 5 6 7 8 9 10
Control for Reconfiguration of UAV
Time (s) Formation”, AIAA Guidance, Navigation, and
Figure 6. Response of the flight path angle of Control Conference, August 11-14, 2003,
UAV approaching aircraft carrier Austin, Tex.