I have been flying RC Planes from the age of 10 and using autopilots from the early Ardupilot of 2009-2010 vintage till the more recent ones. This included the cheap clones that appeared over the last 5-6 years. I have been coding applications for drones since 2009.
My Eureka moment was when flying a particularly unstable aircraft with my remote control on a Sunday morning. I realized that I was employing two types of processes.
Reflex: I was reacting spontaneously to correcting the flight path of the drone against the wind. I really did not think when executing these actions.
Planned moves: This involved flying a circuit, planning the take-off into headwind, speed control on a headwind / tailwind, planning the landing approach, etc. These were processed information at a mental level.
I realized none of the autopilots did this so drones could not be user friendly, APP Expandability was pointless.
Reflex based Controls: Human Split second involuntary reflex a response to external ambient stimuli by the Pilot is key maintaining the aircraft's stability and prevent it from crashing.
Reflex Electronic Controls: ZUPPA KCU: Kinematic Control Unit (Core 1) ZUPPA by responding directly to sensory inputs from multiple sensors at very high control speeds ensures drones stability even under very high wind conditions.
Informative and thought driven Controls: Human While flying an aircraft the Pilot uses the slower Brain power processed information system to guide the aircraft in flight to perform specific tasks like turning, route planning etc.
User input driven Controls: ZUPPA (higher processors can be used) IHOU: Interface and Higher Operations Unit (Core2) handles all higher operations like navigation, APP expansion, etc. This core handles all the functions that make the drone a useful gadget.
Be amongst the select 100 testers to get their hands on the SDK
Building drone apps have never been easier.
Zuppa Application Interface Platform
Smart drones thanks to ZAIP
It’s for JAVA Developers
ZAIP SDK is Open Source