Enable touch support for Android App on Echo Show 15

I have developed an Android App that has full touch support on any standard Android Device. Also the (simulated) touch input on an Android TV Virtual Device works without flaws.
After I transferred the App to my Echo Show 15 (Fire OS via LAT I noticed that the App doesn’t receive any touch events and the virtual remote overlay is shown. Connecting a Bloototh mouse didn’t help either - no click events were received.

I enabled/disabled different features in the AndroidManifest.xml like (android.hardware.touchscreen, android.software.leanback, …) nothing helped. I guess that these feature flags have no effect on touch support at all and only serve as a device filter for the Amazon Appstore.

I already found this tutorial Add touch to Fire TV | Amazon Fire TV but it doesn’t help me because my App doesn’t use the native Android GUI elements and input event handler. Instead it uses the Qt Framework. I don’t know exactly how the Android platform abstraction of Qt works but I think only a SurfaceView is created where the GUI is rendered on via OpenGL ES 2.0. The touch events (and input events in general) get intercepted in the Activity class and get passed through to the Qt event loop.

The question is, how does the Echo Show 15 (or Fire OS in general) recognise that the App has touch support? I assume this is decided during the runtime of the App (and not during compile time). But what technical criteria must be met? If only I knew them, I could patch Qt and maybe get it to work. Unfortunately, the ADB interface has been disabled on the Echo Show 15. So there is no way to examine logcat to find a clue as to what is preventing touch input in my App.

Btw is there even a single App/Game in the Amazon Fire TV Appstore that has full touch support? I haven’t found a single one.

I created a minimal App. Still no touch support. What am I missing?

package com.github.tereius.habview;

import android.os.Bundle;
import android.app.Activity;
import android.view.View;

import android.widget.TextView;
import android.widget.Button;

public class MainActivity extends Activity
    public void onCreate(Bundle savedInstanceState)

        TextView button = new Button(this);
        button.setText("Hello world!");

        button.setOnClickListener(new View.OnClickListener() {
                 public void onClick(View v) {
                      button.setText("Hello again");


    protected void onDestroy()
<?xml version="1.0"?>
<manifest xmlns:android="http://schemas.android.com/apk/res/android"

    <uses-permission android:name="android.permission.INTERNET"/>
    <uses-permission android:name="android.permission.WRITE_EXTERNAL_STORAGE"/>
    <uses-permission android:name="android.permission.ACCESS_NETWORK_STATE"/>
    <uses-permission android:name="android.permission.BLUETOOTH"/>

    <uses-feature android:name="android.hardware.touchscreen" android:required="false"/>
    <uses-feature android:name="android.software.leanback" android:required="false"/>

    <application android:supportsRtl="true">
                <action android:name="android.intent.action.MAIN"/>
                <category android:name="android.intent.category.LAUNCHER"/>

If I remove the bluetooth permission <uses-permission android:name="android.permission.BLUETOOTH"/> from the Manifest I can enable the “touch capabilities” in the LAT.

Those two lines are not shown if the bluetooth permission is declared in the Manifest. Is this a bug in the Appstore?

My last attempt: I enabled “touch capability” in the LAT, compiled the official sample touch app and added the APK to the LAT. Updated the app on my Echo Show 15. Nevertheless, I cannot use touch input and the virtual remote control is displayed.