├── screenshots
├── App_general.jpg
└── App_conf_tab1.png
├── assets
└── deeplabv3_257_mv_gpu.tflite
├── qml.qrc
├── qtquickcontrols2.conf
├── README.md
├── cpp
├── colormanager.h
├── get_top_n.h
├── tensorflowthread.h
├── auxutils.h
├── colormanager.cpp
├── tensorflowthread.cpp
├── objectsrecogfilter.h
├── tensorflow.h
├── objectsrecogfilter.cpp
├── auxutils.cpp
└── tensorflow.cpp
├── storage.js
├── main.cpp
├── Pi_Qt_TFLite_DeepLab.pro
├── Home.qml
├── main.qml
├── Configuration.qml
├── LICENSE
└── Pi_Qt_TFLite_DeepLab.pro.user.4.9-pre1
/screenshots/App_general.jpg:
--------------------------------------------------------------------------------
https://raw.githubusercontent.com/MechatronicsBlog/RaspberryPi_TFLite_DeepLab_Qt/HEAD/screenshots/App_general.jpg
--------------------------------------------------------------------------------
/screenshots/App_conf_tab1.png:
--------------------------------------------------------------------------------
https://raw.githubusercontent.com/MechatronicsBlog/RaspberryPi_TFLite_DeepLab_Qt/HEAD/screenshots/App_conf_tab1.png
--------------------------------------------------------------------------------
/assets/deeplabv3_257_mv_gpu.tflite:
--------------------------------------------------------------------------------
https://raw.githubusercontent.com/MechatronicsBlog/RaspberryPi_TFLite_DeepLab_Qt/HEAD/assets/deeplabv3_257_mv_gpu.tflite
--------------------------------------------------------------------------------
/qml.qrc:
--------------------------------------------------------------------------------
1 |
2 |
3 | main.qml
4 | Home.qml
5 | Configuration.qml
6 | qtquickcontrols2.conf
7 | storage.js
8 |
9 |
10 |
--------------------------------------------------------------------------------
/qtquickcontrols2.conf:
--------------------------------------------------------------------------------
1 | ; This file can be edited to change the style of the application
2 | ; Read "Qt Quick Controls 2 Configuration File" for details:
3 | ; http://doc.qt.io/qt-5/qtquickcontrols2-configuration.html
4 |
5 | [Controls]
6 | Style=Default
7 |
8 | ;[Controls]
9 | ;Style=Material
10 |
11 | ;[Universal]
12 | ;Theme=System
13 | ;Accent=Red
14 |
15 | ;[Material]
16 | ;Theme=Light
17 | ;Accent=Teal
18 | ;Primary=BlueGrey
19 |
--------------------------------------------------------------------------------
/README.md:
--------------------------------------------------------------------------------
1 | # Raspberry Pi - TensorFlow Lite - DeepLab - Qt/QML
2 |
3 | This tutorial shows how to use DeepLab together with TensorFlow Lite and Qt/QML for Raspberry Pi on-device image segmentation.
4 |
5 | Tutorial: https://mechatronicsblog.com/raspberry-pi,-tensorflow-lite-and-qt-qml:-image-segmentation-example/
6 |
7 | App in action: https://youtu.be/SWX_TEOt8B0
8 |
9 |
10 |
11 |
--------------------------------------------------------------------------------
/cpp/colormanager.h:
--------------------------------------------------------------------------------
1 | #ifndef COLORMANAGER_H
2 | #define COLORMANAGER_H
3 |
4 | #include
5 | #include
6 | #include
7 | #include
8 | #include
9 |
10 | class ColorManager
11 | {
12 | public:
13 | QColor getColor(QString element);
14 | static QImage billinearInterpolation(QImage mask, double newHeight, double newWidth);
15 | static QImage applyTransformation(QImage image, QTransform painterTransform);
16 | bool getRgb() const;
17 | void setRgb(bool value);
18 |
19 | private:
20 | QStringList elements;
21 | QList colors;
22 | QColor getNewColor();
23 | bool rgb = true;
24 |
25 | // NOTE: change or add new colors
26 | const QList defColors = {"#f6a625","#99ca53","#2097d2","#b5563d","#7264d6"};
27 | };
28 |
29 | #endif // COLORMANAGER_H
30 |
--------------------------------------------------------------------------------
/storage.js:
--------------------------------------------------------------------------------
1 | function getDatabase() {
2 | return LocalStorage.openDatabaseSync("HelioPi", "0.1", "SettingsDatabase", 100);
3 | }
4 |
5 | function set(setting, value) {
6 | var db = getDatabase();
7 | var res = "";
8 | db.transaction(function(tx) {
9 | tx.executeSql('CREATE TABLE IF NOT EXISTS settings(setting TEXT UNIQUE, value TEXT)');
10 | var rs = tx.executeSql('INSERT OR REPLACE INTO settings VALUES (?,?);', [setting,value]);
11 | if (rs.rowsAffected > 0) {
12 | res = "OK";
13 | } else {
14 | res = "Error";
15 | }
16 | }
17 | );
18 | return res;
19 | }
20 |
21 | function get(setting, default_value) {
22 | var db = getDatabase();
23 | var res="";
24 | try {
25 | db.transaction(function(tx) {
26 | var rs = tx.executeSql('SELECT value FROM settings WHERE setting=?;', [setting]);
27 | if (rs.rows.length > 0) {
28 | res = rs.rows.item(0).value;
29 | } else {
30 | res = default_value;
31 | }
32 | })
33 | } catch (err) {
34 | //console.log("Database " + err);
35 | res = default_value;
36 | };
37 | return res
38 | }
39 |
--------------------------------------------------------------------------------
/main.cpp:
--------------------------------------------------------------------------------
1 | #include
2 | #include
3 | #include
4 |
5 | #include
6 | #include
7 | #include
8 |
9 | #include "tensorflow.h"
10 | #include "auxutils.h"
11 | #include "objectsrecogfilter.h"
12 |
13 | double AuxUtils::angleHor = 0;
14 | double AuxUtils::angleVer = 0;
15 | int AuxUtils::width = 0;
16 | int AuxUtils::height = 0;
17 |
18 | using namespace tflite;
19 |
20 | int main(int argc, char *argv[])
21 | {
22 | QCoreApplication::setAttribute(Qt::AA_EnableHighDpiScaling);
23 |
24 | QGuiApplication app(argc, argv);
25 | app.setOrganizationName("Mechatronics Blog");
26 | app.setOrganizationDomain("mechatronicsblog.com");
27 | app.setApplicationName("TFLite_Qt_Pi");
28 |
29 | QQmlApplicationEngine engine;
30 |
31 | // Register C++ QML types
32 | qmlRegisterType("TensorFlow", 1, 0, "TensorFlow");
33 | qmlRegisterType("ObjectsRecognizer", 1, 0, "ObjectsRecognizer");
34 | qmlRegisterType("AuxUtils", 1, 0, "AuxUtils");
35 |
36 | // Global objects
37 | AuxUtils* auxUtils = new AuxUtils();
38 | engine.rootContext()->setContextProperty("auxUtils",auxUtils);
39 | engine.rootContext()->setContextProperty("globalEngine",&engine);
40 |
41 | // Register meta types
42 | qRegisterMetaType>("QList");
43 | qRegisterMetaType>("QList");
44 |
45 | engine.load(QUrl(QStringLiteral("qrc:/main.qml")));
46 | if (engine.rootObjects().isEmpty())
47 | return -1;
48 |
49 | return app.exec();
50 | }
51 |
--------------------------------------------------------------------------------
/cpp/get_top_n.h:
--------------------------------------------------------------------------------
1 | /* Copyright 2017 The TensorFlow Authors. All Rights Reserved.
2 |
3 | Licensed under the Apache License, Version 2.0 (the "License");
4 | you may not use this file except in compliance with the License.
5 | You may obtain a copy of the License at
6 |
7 | http://www.apache.org/licenses/LICENSE-2.0
8 |
9 | Unless required by applicable law or agreed to in writing, software
10 | distributed under the License is distributed on an "AS IS" BASIS,
11 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 | See the License for the specific language governing permissions and
13 | limitations under the License.
14 | ==============================================================================*/
15 |
16 | #ifndef TENSORFLOW_CONTRIB_LITE_EXAMPLES_LABEL_IMAGE_GET_TOP_N_H
17 | #define TENSORFLOW_CONTRIB_LITE_EXAMPLES_LABEL_IMAGE_GET_TOP_N_H
18 |
19 | #include "tensorflow/lite/examples/label_image/get_top_n_impl.h"
20 |
21 | namespace tflite {
22 | namespace label_image {
23 |
24 | template
25 | void get_top_n(T* prediction, int prediction_size, size_t num_results,
26 | float threshold, std::vector>* top_results,
27 | bool input_floating);
28 |
29 | // explicit instantiation so that we can use them otherwhere
30 | template void get_top_n(uint8_t*, int, size_t, float,
31 | std::vector>*, bool);
32 | template void get_top_n(float*, int, size_t, float,
33 | std::vector>*, bool);
34 |
35 | } // namespace label_image
36 | } // namespace tflite
37 |
38 | #endif // TENSORFLOW_CONTRIB_LITE_EXAMPLES_LABEL_IMAGE_GET_TOP_N_H
39 |
--------------------------------------------------------------------------------
/cpp/tensorflowthread.h:
--------------------------------------------------------------------------------
1 | #ifndef TENSORFLOWTHREAD_H
2 | #define TENSORFLOWTHREAD_H
3 |
4 | #include
5 | #include
6 | #include
7 | #include
8 | #include
9 | #include "tensorflow.h"
10 |
11 | class WorkerTF: public QObject
12 | {
13 | Q_OBJECT
14 |
15 | QImage imgTF;
16 | TensorFlow *tf;
17 | QString source;
18 | QString destination;
19 | bool videoMode;
20 | QMap activeLabels;
21 | bool showAim;
22 | bool showInfTime;
23 |
24 | public:
25 | void setImgTF(const QImage &value);
26 | void setTf(TensorFlow *value);
27 | void setVideoInfo(QString s, QString d, bool sAim, bool sInfTime, QMap aLabels);
28 |
29 | public slots:
30 | void work();
31 |
32 | private:
33 | QImage processImage(QImage img);
34 |
35 | signals:
36 | void results(int network, QStringList captions, QList confidences, QList boxes, QList masks, int infTime);
37 | void finished();
38 | void numFrame(int n);
39 | void numFrames(int n);
40 | };
41 |
42 | class TensorFlowThread: public QObject
43 | {
44 | Q_OBJECT
45 |
46 | public:
47 | TensorFlowThread();
48 | void setTf(TensorFlow *value);
49 |
50 | void run(QImage imgTF);
51 | void run(QString source, QString destination, bool showAim, bool showInfTime, QMap activeLabels);
52 |
53 | signals:
54 | void results(int network, QStringList captions, QList confidences, QList boxes, QList masks, int infTime);
55 | void numFrame(int n);
56 | void numFrames(int n);
57 |
58 | public slots:
59 | void propagateResults(int network, QStringList captions, QList confidences, QList boxes, QList masks, int infTime);
60 | void propagateNumFrame(int n);
61 | void propagateNumFrames(int n);
62 |
63 | private:
64 | QThread threadTF;
65 | WorkerTF worker;
66 | };
67 |
68 | #endif // TENSORFLOWTHREAD_H
69 |
--------------------------------------------------------------------------------
/Pi_Qt_TFLite_DeepLab.pro:
--------------------------------------------------------------------------------
1 | QT += quick multimedia sensors multimedia-private
2 | CONFIG += console c++11 qml_debug
3 |
4 | # The following define makes your compiler emit warnings if you use
5 | # any feature of Qt which as been marked deprecated (the exact warnings
6 | # depend on your compiler). Please consult the documentation of the
7 | # deprecated API in order to know how to port your code away from it.
8 | DEFINES += QT_DEPRECATED_WARNINGS
9 |
10 | # You can also make your code fail to compile if you use deprecated APIs.
11 | # In order to do so, uncomment the following line.
12 | # You can also select to disable deprecated APIs only up to a certain version of Qt.
13 | #DEFINES += QT_DISABLE_DEPRECATED_BEFORE=0x060000 # disables all the APIs deprecated before Qt 6.0.0
14 |
15 | RESOURCES += qml.qrc
16 |
17 | # Additional import path used to resolve QML modules in Qt Creator's code model
18 | QML_IMPORT_PATH =
19 |
20 | # Additional import path used to resolve QML modules just for Qt Quick Designer
21 | QML_DESIGNER_IMPORT_PATH =
22 |
23 | INCLUDEPATH += $$PWD/cpp \
24 | $$PWD/../tensorflow \
25 | $$PWD/../tensorflow/tensorflow/lite/tools/make/downloads/flatbuffers/include
26 |
27 | # We consider Linux and distinguish between Raspbian (for Raspberry Pi) and other Linux distributions
28 | linux{
29 | contains(QMAKE_CXX, .*raspbian.*arm.*):{
30 | # TensorFlow Lite lib path
31 | LIBS += -L$$PWD/../tensorflow/tensorflow/lite/tools/make/gen/rpi_armv7l/lib
32 |
33 | # Assets to be deployed: path and files
34 | assets.path = /home/pi/qt_apps/$${TARGET}/bin/assets
35 | assets.files = assets/*
36 | }
37 | else {
38 | # TensorFlow Lite lib path
39 | LIBS += -L$$PWD/../tensorflow/tensorflow/lite/tools/make/gen/linux_x86_64/lib
40 |
41 | # Assets to be deployed: path and files
42 | # WARNING: Define yourself the path!
43 | # assets.path = /home/user/app
44 | assets.files = assets/*
45 | }
46 | }
47 |
48 | LIBS += -ltensorflow-lite -ldl
49 |
50 | # Default rules for deployment.
51 | qnx: target.path = /tmp/$${TARGET}/bin
52 | else: unix:!android: target.path = /home/pi/qt_apps/$${TARGET}/bin
53 | !isEmpty(target.path): INSTALLS += target
54 | INSTALLS += assets
55 |
56 | HEADERS += \
57 | cpp/objectsrecogfilter.h \
58 | cpp/tensorflow.h \
59 | cpp/tensorflowthread.h \
60 | cpp/colormanager.h \
61 | cpp/get_top_n.h \
62 | cpp/auxutils.h
63 |
64 | SOURCES += \
65 | main.cpp \
66 | cpp/objectsrecogfilter.cpp \
67 | cpp/tensorflow.cpp \
68 | cpp/tensorflowthread.cpp \
69 | cpp/colormanager.cpp \
70 | cpp/auxutils.cpp
71 |
72 | DISTFILES += \
73 | Home.qml \
74 | Configuration.qml \
75 | main.qml
76 |
--------------------------------------------------------------------------------
/cpp/auxutils.h:
--------------------------------------------------------------------------------
1 | #ifndef UTILS_H
2 | #define UTILS_H
3 |
4 | #include
5 | #include
6 | #include
7 | #include
8 | #include
9 | #include
10 | #include
11 |
12 | const QString assetsPath = "./assets";
13 | const QString modelName = "deeplabv3_257_mv_gpu.tflite";
14 | const QString labelsName = "labelmap.txt";
15 | const QString RES_CHAR = "x";
16 |
17 | class AuxUtils : public QObject
18 | {
19 | Q_OBJECT
20 |
21 | public slots:
22 | static QImage drawText(QImage image, QRectF rect, QString text, Qt::AlignmentFlag pos = Qt::AlignBottom,
23 | Qt::GlobalColor borderColor = Qt::black,
24 | double borderSize = 0.5,
25 | Qt::GlobalColor fontColor = Qt::white,
26 | QFont font = QFont("Times", 16, QFont::Bold));
27 | static QImage drawBoxes(QImage image, QRect rect, QStringList captions, QList confidences, QList boxes, double minConfidence,
28 | QMap activeLabels, bool rgb);
29 | static QImage drawMasks(QImage image, QRect rect, QStringList captions, QList confidences, QList boxes, QList masks, double minConfidence, QMap activeLabels);
30 | static QImage drawSegmentation(QImage image, QImage segmentation);
31 | static QString getDefaultModelFilename();
32 | static QString getDefaultLabelsFilename();
33 | static QRectF frameMatchImg(QImage img, QSize rectSize);
34 | static int sp(int pixel, QSizeF size);
35 | static double dpi(QSizeF size);
36 | static QString deviceInfo();
37 | static QString qtVersion();
38 | static QString getAssetsPath();
39 | static QImage setOpacity(QImage& image, qreal opacity);
40 | static bool isBGRvideoFrame(QVideoFrame f);
41 | static bool isBGRimage(QImage i);
42 | bool readLabels(QString filename);
43 | QStringList getLabels();
44 | int numberThreads();
45 | static QVariantList networkInterfaces();
46 | static void setAngleHor(double angle);
47 | static void setAngleVer(double angle);
48 | static bool setResolution(QString res);
49 |
50 | signals:
51 | void imageSaved(QString file);
52 |
53 | private:
54 | static QString copyIfNotExistOrUpdate(QString file, QString defFile);
55 | static QByteArray fileMD5(QString filename);
56 |
57 | // Constant values
58 | static constexpr int FONT_PIXEL_SIZE_TEXT = 38;
59 | static constexpr int FONT_PIXEL_SIZE_BOX = 24;
60 | static constexpr double MASK_OPACITY = 0.6;
61 | static constexpr double LINE_WIDTH = 2;
62 | static constexpr int FONT_HEIGHT_MARGIN = 3;
63 | static constexpr int FONT_WIDTH_MARGIN = 6;
64 | QStringList labels;
65 |
66 | public:
67 | static double angleHor;
68 | static double angleVer;
69 | static int width;
70 | static int height;
71 |
72 | };
73 |
74 | #endif // UTILS_H
75 |
--------------------------------------------------------------------------------
/cpp/colormanager.cpp:
--------------------------------------------------------------------------------
1 | #include "colormanager.h"
2 |
3 | #include
4 |
5 | QColor ColorManager::getColor(QString element)
6 | {
7 | int index = elements.indexOf(element);
8 |
9 | if (index>=0) return colors.at(index);
10 |
11 | QColor newColor = getNewColor();
12 | elements.append(element);
13 | colors.append(newColor);
14 | return newColor;
15 | }
16 |
17 | QColor ColorManager::getNewColor()
18 | {
19 | QColor color = defColors.at(elements.count()%defColors.count()).toRgb();
20 |
21 | if (!rgb)
22 | {
23 | int r = color.red();
24 | int b = color.blue();
25 |
26 | color.setRed(b);
27 | color.setBlue(r);
28 | }
29 |
30 | return color;
31 | }
32 |
33 | bool ColorManager::getRgb() const
34 | {
35 | return rgb;
36 | }
37 |
38 | void ColorManager::setRgb(bool value)
39 | {
40 | rgb = value;
41 | }
42 |
43 | int getColor(QImage mask, QColor color, int x, int y)
44 | {
45 | return color == Qt::red ? qRed(mask.pixel(x,y)) :
46 | color == Qt::blue ? qBlue(mask.pixel(x,y)) :
47 | qGreen(mask.pixel(x,y));
48 | }
49 |
50 | int billinerColor(QImage mask, QColor color, int xa, int xb, int xc, int xd, int ya, int yb, int yc, int yd, double alpha, double beta)
51 | {
52 | int pa,pb,pc,pd;
53 |
54 | pa = getColor(mask,color,xa,ya);
55 | pb = getColor(mask,color,xb,yb);
56 | pc = getColor(mask,color,xc,yc);
57 | pd = getColor(mask,color,xd,yd);
58 | return (1-alpha)*(1-beta)*pa+alpha*(1-beta)*pb+
59 | (1-alpha)*beta*pc + alpha*beta*pd + 0.5;
60 | }
61 |
62 | uint billinearPixel(QImage mask, double sx, double sy, int k, int j)
63 | {
64 | double alpha,beta;
65 | int xa,xb,xc,xd,ya,yb,yc,yd;
66 |
67 | xa = k/sx; ya = j/sy;
68 | xb = xa+1; yb = ya;
69 | xc = xa; yc = ya+1;
70 | xd = xa+1; yd = ya+1;
71 | if (xb>=mask.width()) xb--;
72 | if (xd>=mask.width()) xd--;
73 | if (yc>=mask.height()) yc--;
74 | if (yd>=mask.height()) yd--;
75 | alpha = k/sx - xa;
76 | beta = j/sy - ya;
77 |
78 | int red = billinerColor(mask,Qt::red,xa,xb,xc,xd,ya,yb,yc,yd,alpha,beta);
79 | int green = billinerColor(mask,Qt::green,xa,xb,xc,xd,ya,yb,yc,yd,alpha,beta);
80 | int blue = billinerColor(mask,Qt::blue,xa,xb,xc,xd,ya,yb,yc,yd,alpha,beta);
81 |
82 | return qRgb(red,green,blue);
83 | }
84 |
85 | QImage ColorManager::billinearInterpolation(QImage mask, double newHeight, double newWidth)
86 | {
87 | const double sy = newHeight/mask.height();
88 | const double sx = newWidth/mask.width();
89 |
90 | // Resize mask to box size
91 | QImage maskScaled(mask.width()*sx,mask.height()*sy,QImage::Format_ARGB32_Premultiplied);
92 | maskScaled.fill(Qt::transparent);
93 |
94 | // Billinear interpolation
95 | // https://chu24688.tian.yam.com/posts/44797337
96 | for(int j=0;j aLabels)
16 | {
17 | source = s;
18 | destination = d;
19 | showAim = sAim;
20 | showInfTime = sInfTime;
21 | activeLabels = aLabels;
22 | videoMode = true;
23 | }
24 |
25 | void WorkerTF::work()
26 | {
27 | if (!videoMode)
28 | {
29 | tf->run(imgTF);
30 | emit finished();
31 | emit results(tf->getKindNetwork(),tf->getResults(),tf->getConfidence(),tf->getBoxes(),tf->getMasks(),tf->getInferenceTime());
32 | }
33 | }
34 |
35 | QImage WorkerTF::processImage(QImage img)
36 | {
37 | // Data
38 | double minConf = tf->getThreshold();
39 | int inferenceTime = tf->getInferenceTime();
40 | QStringList results = tf->getResults();
41 | QList confidence = tf->getConfidence();
42 | QList boxes = tf->getBoxes();
43 | QList masks = tf->getMasks();
44 |
45 | // Draw masks on image
46 | if (!masks.isEmpty())
47 | img = AuxUtils::drawMasks(img,img.rect(),results,confidence,boxes,masks,minConf,activeLabels);
48 |
49 | // Draw boxes on image
50 | img = AuxUtils::drawBoxes(img,img.rect(),results,confidence,boxes,minConf,activeLabels,true);
51 |
52 | // Show inference time
53 | if (showInfTime)
54 | {
55 | QString text = QString::number(inferenceTime) + " ms";
56 | img = AuxUtils::drawText(img,img.rect(),text);
57 | }
58 |
59 | return img;
60 | }
61 |
62 | TensorFlowThread::TensorFlowThread()
63 | {
64 | threadTF.setObjectName("TensorFlow thread");
65 | worker.moveToThread(&threadTF);
66 | QObject::connect(&worker, SIGNAL(results(int, QStringList, QList, QList, QList, int)), this, SLOT(propagateResults(int, QStringList, QList, QList, QList, int)));
67 | QObject::connect(&worker, SIGNAL(numFrame(int)), this, SLOT(propagateNumFrame(int)));
68 | QObject::connect(&worker, SIGNAL(numFrames(int)), this, SLOT(propagateNumFrames(int)));
69 | QObject::connect(&worker, SIGNAL(finished()), &threadTF, SLOT(quit()));
70 | QObject::connect(&threadTF, SIGNAL(started()), &worker, SLOT(work()));
71 | }
72 |
73 | void TensorFlowThread::setTf(TensorFlow *value)
74 | {
75 | worker.setTf(value);
76 | }
77 |
78 | void TensorFlowThread::run(QImage imgTF)
79 | {
80 | worker.setImgTF(imgTF);
81 | threadTF.start();
82 | }
83 |
84 | void TensorFlowThread::run(QString source, QString destination, bool showAim, bool showInfTime, QMap activeLabels)
85 | {
86 | worker.setVideoInfo(source,destination,showAim,showInfTime,activeLabels);
87 | threadTF.start();
88 | }
89 |
90 | void TensorFlowThread::propagateResults(int network, QStringList captions, QList confidences, QList boxes, QList masks, int infTime)
91 | {
92 | emit results(network,captions,confidences,boxes,masks,infTime);
93 | }
94 |
95 | void TensorFlowThread::propagateNumFrame(int n)
96 | {
97 | emit numFrame(n);
98 | }
99 |
100 | void TensorFlowThread::propagateNumFrames(int n)
101 | {
102 | emit numFrames(n);
103 | }
104 |
--------------------------------------------------------------------------------
/Home.qml:
--------------------------------------------------------------------------------
1 | import QtQuick 2.9
2 | import QtQuick.Controls 2.2
3 | import QtMultimedia 5.11
4 |
5 | import ObjectsRecognizer 1.0
6 | import AuxUtils 1.0
7 |
8 | Page {
9 | id: root
10 | title: qsTr("Live")
11 |
12 | readonly property double fontPixelSize: Qt.application.font.pixelSize * 1.6
13 | readonly property var defResolutions: ["320x240","640x480","800x480","800x600","1024x768","1280x720","1920x1080"]
14 |
15 | property double minConfidence
16 | property int nThreads
17 | property bool showInfTime
18 | property string resolution
19 | property var resolutions: []
20 | property bool semiTransparent
21 | property bool showBackground
22 |
23 | onResolutionChanged: auxUtils.setResolution(resolution)
24 |
25 | background: Rectangle { color: 'black'}
26 |
27 | signal cameraResolutionsLoaded()
28 |
29 | Camera {
30 | id: camera
31 | deviceId: QtMultimedia.defaultCamera.deviceId
32 | captureMode: Camera.CaptureViewfinder
33 | property bool availableCamera: QtMultimedia.availableCameras.length>0
34 | viewfinder.resolution: resolution
35 |
36 | onCameraStateChanged: {
37 | if (camera.cameraState === 2)
38 | {
39 | var res = camera.supportedViewfinderResolutions()
40 |
41 | if (res.length>0)
42 | {
43 | resolutions = []
44 |
45 | for(var i=0; i 1 ? "\u25C0" : "\u2630"
39 | font.pixelSize: Qt.application.font.pixelSize * 1.6
40 | onClicked: {
41 | if (stackView.depth > 1) {
42 | stackView.pop()
43 | } else {
44 | drawer.open()
45 | }
46 | }
47 | }
48 |
49 | Label {
50 | text: stackView.currentItem.title
51 | anchors.centerIn: parent
52 | }
53 | }
54 |
55 | Drawer {
56 | id: drawer
57 | width: window.width * 0.3
58 | height: window.height
59 |
60 | Column {
61 | anchors.fill: parent
62 |
63 | ItemDelegate {
64 | text: qsTr("Settings")
65 | width: parent.width
66 | onClicked: {
67 | stackView.push(configuration)
68 | drawer.close()
69 | }
70 | }
71 | }
72 | }
73 |
74 | StackView {
75 | id: stackView
76 | initialItem: home
77 | anchors.fill: parent
78 | }
79 |
80 | Home{
81 | id: home
82 | visible: false
83 | minConfidence: window.minConfidence
84 | nThreads: window.nThreads
85 | showInfTime: window.showInfTime
86 | resolution: window.resolution
87 | semiTransparent: window.semiTransparent
88 | showBackground: window.showBackground
89 | }
90 |
91 | Configuration{
92 | id: configuration
93 | visible: false
94 |
95 | minConfidence: window.minConfidence
96 | nThreads: window.nThreads
97 | showInfTime: window.showInfTime
98 | resolution: window.resolution
99 | resolutions: home.resolutions
100 | semiTransparent: window.semiTransparent
101 | showBackground: window.showBackground
102 |
103 | onMinConfidenceChanged: {Settings.set(tMinConfidence,minConfidence); window.minConfidence = minConfidence }
104 | onNThreadsChanged: {Settings.set(tNThreads,nThreads); window.nThreads = nThreads }
105 | onShowInfTimeChanged: {Settings.set(tShowInfTime,showInfTime); window.showInfTime = showInfTime }
106 | onResolutionUpdated: {Settings.set(tResolution,resolution); window.resolution = resolution}
107 | onSemiTransparentChanged: {Settings.set(tSemiTransparent,semiTransparent); window.semiTransparent = semiTransparent}
108 | onShowBackgroundChanged: {Settings.set(tShowBackground,showBackground); window.showBackground = showBackground}
109 | }
110 | }
111 |
--------------------------------------------------------------------------------
/cpp/objectsrecogfilter.h:
--------------------------------------------------------------------------------
1 | #ifndef OBJECTSRECOGFILTER_H
2 | #define OBJECTSRECOGFILTER_H
3 |
4 | #include
5 | #include
6 | #include
7 | #include
8 | #include
9 |
10 | #include "tensorflow.h"
11 | #include "tensorflowthread.h"
12 |
13 | // https://stackoverflow.com/questions/43106069/how-to-convert-qvideoframe-with-yuv-data-to-qvideoframe-with-rgba32-data-in
14 |
15 | class ObjectsRecogFilterRunable;
16 |
17 | class ObjectsRecogFilter : public QAbstractVideoFilter
18 | {
19 | Q_OBJECT
20 |
21 | Q_PROPERTY(double cameraOrientation READ getCameraOrientation WRITE setCameraOrientation)
22 | Q_PROPERTY(double videoOrientation READ getVideoOrientation WRITE setVideoOrientation)
23 | Q_PROPERTY(double minConfidence READ getMinConfidence WRITE setMinConfidence)
24 | Q_PROPERTY(QSize contentSize READ getContentSize WRITE setContentSize)
25 | Q_PROPERTY(bool acceleration READ getAcceleration WRITE setAcceleration)
26 | Q_PROPERTY(int nThreads READ getNThreads WRITE setNThreads)
27 | Q_PROPERTY(bool showInfTime READ getShowInfTime WRITE setShowInfTime)
28 | Q_PROPERTY(double angle READ getAngle NOTIFY angleChanged)
29 | Q_PROPERTY(bool semiTransparent READ getSemiTransparent WRITE setSemiTransparent)
30 | Q_PROPERTY(bool showBackground READ getShowBackground WRITE setShowBackground)
31 |
32 | public slots:
33 | void init(int imgHeight, int imgWidth);
34 | void initInput(int imgHeight, int imgWidth);
35 | QMap getActiveLabels();
36 | bool getActiveLabel(QString key);
37 | void setActiveLabel(QString key, bool value);
38 |
39 | public:
40 | ObjectsRecogFilter();
41 | QVideoFilterRunnable *createFilterRunnable();
42 | void setCameraOrientation(double o);
43 | void setVideoOrientation(double o);
44 | double getCameraOrientation();
45 | double getVideoOrientation();
46 | double getMinConfidence() const;
47 | void setMinConfidence(double value);
48 | bool getRunning();
49 | void releaseRunning();
50 | QSize getContentSize() const;
51 | void setContentSize(const QSize &value);
52 | bool getAcceleration() const;
53 | void setAcceleration(bool value);
54 | int getNThreads() const;
55 | void setNThreads(int value);
56 | bool getShowInfTime() const;
57 | void setShowInfTime(bool value);
58 | void setFrameSize(QSize size);
59 | void setFrameRate(int fps);
60 | double getAngle() const;
61 | void setAngle(const double value);
62 | double getImgHeight();
63 | double getImgWidth();
64 | bool getInitialized() const;
65 | void setInitialized(bool value);
66 | bool getSemiTransparent() const;
67 | void setSemiTransparent(bool value);
68 | bool getShowBackground() const;
69 | void setShowBackground(bool value);
70 |
71 | private:
72 | ObjectsRecogFilterRunable *rfr;
73 | TensorFlow tf;
74 | TensorFlowThread tft;
75 | double camOrientation;
76 | double vidOrientation;
77 | double minConf;
78 | bool acc;
79 | int nThr;
80 | bool infTime;
81 | QMutex mutex;
82 | bool running;
83 | QSize videoSize;
84 | QMap activeLabels;
85 | void setRunning(bool val);
86 | double ang;
87 | bool initialized;
88 | bool transparent;
89 | bool background;
90 |
91 | signals:
92 | void runTensorFlow(QImage imgTF);
93 | void focusDataChanged();
94 | void focusedChanged();
95 | void angleChanged();
96 | void errorXChanged();
97 | void errorYChanged();
98 |
99 | private slots:
100 | void TensorFlowExecution(QImage imgTF);
101 | void processResults(int network, QStringList res, QList conf, QList boxes, QList masks, int inftime);
102 | };
103 |
104 | class ObjectsRecogFilterRunable : public QVideoFilterRunnable
105 | {
106 | public:
107 | ObjectsRecogFilterRunable(ObjectsRecogFilter *filter, QStringList res);
108 | QVideoFrame run(QVideoFrame *input, const QVideoSurfaceFormat &surfaceFormat, RunFlags flags);
109 | void setResults(int net, QStringList res, QList conf, QList box, QList mask, int inftime);
110 |
111 | private:
112 | ObjectsRecogFilter *m_filter;
113 | int network;
114 | QStringList results;
115 | QList confidence;
116 | QList boxes;
117 | QList masks;
118 | int inferenceTime;
119 | QElapsedTimer timer;
120 |
121 | };
122 |
123 | #endif // OBJECTSRECOGFILTER_H
124 |
--------------------------------------------------------------------------------
/cpp/tensorflow.h:
--------------------------------------------------------------------------------
1 | #ifndef TENSORFLOW_H
2 | #define TENSORFLOW_H
3 |
4 | #include
5 | #include
6 | #include
7 |
8 | #include "tensorflow/lite/error_reporter.h"
9 | #include "tensorflow/lite/interpreter.h"
10 | #include "tensorflow/lite/model.h"
11 | #include "tensorflow/lite/graph_info.h"
12 | #include "tensorflow/lite/kernels/register.h"
13 |
14 | using namespace tflite;
15 |
16 | class TensorFlow : public QObject
17 | {
18 | Q_OBJECT
19 | public:
20 | explicit TensorFlow(QObject *parent = nullptr);
21 | ~TensorFlow();
22 |
23 | static const int knIMAGE_SEGMENTATION = 1;
24 | static const int knOBJECT_DETECTION = 2;
25 | static const int DEF_BOX_DISTANCE = 10;
26 |
27 | const QStringList DeepLab_names = {"background", "aeroplane", "bicycle", "bird", "boat", "bottle", "bus", "car", "cat", "chair",
28 | "cow", "diningtable", "dog", "horse", "motorbike", "person", "pottedplant", "sheep", "sofa", "train", "tv"};
29 |
30 | public slots:
31 | bool init(int imgHeight, int imgWidth);
32 | bool run(QImage img);
33 | QString getFilename() const;
34 | void setFilename(const QString &value);
35 | QString getLabelsFilename() const;
36 | void setLabelsFilename(const QString &value);
37 | bool getAccelaration() const;
38 | void setAccelaration(bool value);
39 | bool getVerbose() const;
40 | void setVerbose(bool value);
41 | int getNumThreads() const;
42 | void setNumThreads(int value);
43 | int getHeight() const;
44 | int getWidth() const;
45 | int getChannels() const;
46 | QString getLabel(int index);
47 | QString getResultCaption(int index);
48 | double getResultConfidence(int index);
49 | QStringList getResults();
50 | QList getConfidence();
51 | QList getBoxes();
52 | QList getMasks();
53 | int getInferenceTime();
54 | int getKindNetwork();
55 | double getThreshold() const;
56 | void setThreshold(double value);
57 | void initInput(int imgHeight, int imgWidth);
58 | bool initTFLite(int imgHeight, int imgWidth);
59 | bool setInputsTFLite(QImage image);
60 | bool inferenceTFLite();
61 | bool getSegmentationOutputs(QImage &segmentation);
62 | bool getObjectOutputs(QStringList &captions, QList &confidences, QList &locations, QList &images);
63 | bool getShowBackground() const;
64 | void setShowBackground(bool value);
65 | bool getSemiTransparent() const;
66 | void setSemiTransparent(bool value);
67 |
68 | private:
69 | // Configuration constants
70 | const double MASK_THRESHOLD = 0.3;
71 |
72 | // Output names
73 | const QString num_detections = "num_detections";
74 | const QString detection_classes = "detection_classes";
75 | const QString detection_scores = "detection_scores";
76 | const QString detection_boxes = "detection_boxes";
77 | const QString detection_masks = "detection_masks";
78 |
79 | // Segementation
80 | const QColor background_color = Qt::black;
81 | const float object_alpha = 0.5;
82 | const QList objects_colors = {background_color,Qt::gray,Qt::blue,Qt::green,Qt::yellow,Qt::lightGray,Qt::darkBlue,Qt::darkRed,Qt::cyan,Qt::darkYellow,
83 | Qt::darkMagenta,Qt::darkCyan,Qt::white,Qt::darkGreen,Qt::darkGray,Qt::red,Qt::magenta,"#ffaa00","#662c91","#d943b4","#aaaa7f"};
84 |
85 | // Network configuration
86 | bool has_detection_masks;
87 |
88 | // Threshold
89 | double threshold;
90 |
91 | // Segmentation
92 | bool showBackground;
93 | bool semiTransparent;
94 |
95 | // Image properties
96 | const QImage::Format format = QImage::Format_RGB888;
97 | const int numChannels = 3;
98 |
99 | // Kind of network in model
100 | int kind_network;
101 |
102 | // Model filename
103 | QString filename;
104 |
105 | // Labels filename & data
106 | QString labelsFilename;
107 | QStringList labels;
108 |
109 | // Results
110 | QStringList rCaption;
111 | QList rConfidence;
112 | QList rBox;
113 | QList rMasks;
114 | int inferenceTime;
115 |
116 | // Initialized
117 | bool initialized;
118 |
119 | // Accelaration
120 | bool accelaration;
121 |
122 | // Verbose
123 | bool verbose;
124 |
125 | // Number of threads
126 | int numThreads;
127 |
128 | // Image configuration
129 | int wanted_height, wanted_width, wanted_channels;
130 | int img_height, img_width, img_channels;
131 |
132 | // Model
133 | std::unique_ptr model;
134 | // Resolver
135 | tflite::ops::builtin::BuiltinOpResolver resolver;
136 | // Interpreter
137 | std::unique_ptr interpreter;
138 | // Error reporter
139 | StderrReporter error_reporter;
140 | // Outputs
141 | std::vector outputs;
142 |
143 | // Private functions
144 | bool readLabels();
145 | bool setInputs(QImage image);
146 | bool inference();
147 | };
148 |
149 | #endif // TENSORFLOW_H
150 |
--------------------------------------------------------------------------------
/cpp/objectsrecogfilter.cpp:
--------------------------------------------------------------------------------
1 | #include
2 | #include
3 | #include
4 | #include
5 | #include
6 | #include
7 |
8 | #include "auxutils.h"
9 | #include "private/qvideoframe_p.h"
10 |
11 | // WARNING: same TensorFlow initialization repeated in ObjectRecogFilter and TensorFlowQML constructors
12 | ObjectsRecogFilter::ObjectsRecogFilter()
13 | {
14 | connect(this, SIGNAL(runTensorFlow(QImage)), this, SLOT(TensorFlowExecution(QImage)));
15 | connect(&tft,SIGNAL(results(int, QStringList, QList, QList, QList, int)),this,SLOT(processResults(int, QStringList, QList, QList, QList, int)));
16 |
17 | tf.setFilename(AuxUtils::getDefaultModelFilename());
18 | tf.setLabelsFilename(AuxUtils::getDefaultLabelsFilename());
19 | tf.setAccelaration(true);
20 | tf.setNumThreads(QThread::idealThreadCount());
21 |
22 | releaseRunning();
23 | initialized = false;
24 | }
25 |
26 | void ObjectsRecogFilter::init(int imgHeight, int imgWidth)
27 | {
28 | initialized = tf.init(imgHeight,imgWidth);
29 | tft.setTf(&tf);
30 | }
31 |
32 | void ObjectsRecogFilter::initInput(int imgHeight, int imgWidth)
33 | {
34 | tf.initInput(imgHeight,imgWidth);
35 | }
36 |
37 | void ObjectsRecogFilter::TensorFlowExecution(QImage imgTF)
38 | {
39 | tf.setAccelaration(getAcceleration());
40 | tf.setNumThreads(getNThreads());
41 | tft.run(imgTF);
42 | }
43 |
44 | void ObjectsRecogFilter::processResults(int network, QStringList res, QList conf, QList boxes, QList masks, int inftime)
45 | {
46 | rfr->setResults(network,res,conf,boxes,masks,inftime);
47 | releaseRunning();
48 | }
49 |
50 | void ObjectsRecogFilter::setCameraOrientation(double o)
51 | {
52 | camOrientation = o;
53 | }
54 |
55 | void ObjectsRecogFilter::setVideoOrientation(double o)
56 | {
57 | vidOrientation = o;
58 | }
59 |
60 | double ObjectsRecogFilter::getCameraOrientation()
61 | {
62 | return camOrientation;
63 | }
64 |
65 | double ObjectsRecogFilter::getVideoOrientation()
66 | {
67 | return vidOrientation;
68 | }
69 |
70 | bool ObjectsRecogFilter::getRunning()
71 | {
72 | QMutexLocker locker(&mutex);
73 |
74 | bool val = running;
75 | if (!val) setRunning(true);
76 |
77 | return !val;
78 | }
79 |
80 | void ObjectsRecogFilter::setRunning(bool val)
81 | {
82 | running = val;
83 | }
84 |
85 | bool ObjectsRecogFilter::getShowBackground() const
86 | {
87 | return background;
88 | }
89 |
90 | void ObjectsRecogFilter::setShowBackground(bool value)
91 | {
92 | background = value;
93 | tf.setShowBackground(background);
94 | }
95 |
96 | bool ObjectsRecogFilter::getSemiTransparent() const
97 | {
98 | return transparent;
99 | }
100 |
101 | void ObjectsRecogFilter::setSemiTransparent(bool value)
102 | {
103 | transparent = value;
104 | tf.setSemiTransparent(transparent);
105 | }
106 |
107 | bool ObjectsRecogFilter::getInitialized() const
108 | {
109 | return initialized;
110 | }
111 |
112 | void ObjectsRecogFilter::setInitialized(bool value)
113 | {
114 | initialized = value;
115 | }
116 |
117 | void ObjectsRecogFilter::releaseRunning()
118 | {
119 | QMutexLocker locker(&mutex);
120 |
121 | setRunning(false);
122 | }
123 |
124 | QSize ObjectsRecogFilter::getContentSize() const
125 | {
126 | return videoSize;
127 | }
128 |
129 | void ObjectsRecogFilter::setContentSize(const QSize &value)
130 | {
131 | videoSize = value;
132 | }
133 |
134 | bool ObjectsRecogFilter::getAcceleration() const
135 | {
136 | return acc;
137 | }
138 |
139 | void ObjectsRecogFilter::setAcceleration(bool value)
140 | {
141 | acc = value;
142 | }
143 |
144 | int ObjectsRecogFilter::getNThreads() const
145 | {
146 | return nThr;
147 | }
148 |
149 | void ObjectsRecogFilter::setNThreads(int value)
150 | {
151 | nThr = value;
152 | }
153 |
154 | bool ObjectsRecogFilter::getShowInfTime() const
155 | {
156 | return infTime;
157 | }
158 |
159 | void ObjectsRecogFilter::setShowInfTime(bool value)
160 | {
161 | infTime = value;
162 | }
163 |
164 | double ObjectsRecogFilter::getMinConfidence() const
165 | {
166 | return minConf;
167 | }
168 |
169 | void ObjectsRecogFilter::setMinConfidence(double value)
170 | {
171 | minConf = value;
172 | tf.setThreshold(minConf);
173 | }
174 |
175 | ObjectsRecogFilterRunable::ObjectsRecogFilterRunable(ObjectsRecogFilter *filter, QStringList res)
176 | {
177 | m_filter = filter;
178 | results = res;
179 | }
180 |
181 | void ObjectsRecogFilterRunable::setResults(int net, QStringList res, QList conf, QList box, QList mask, int inftime)
182 | {
183 | network = net;
184 | results = res;
185 | confidence = conf;
186 | boxes = box;
187 | masks = mask;
188 | inferenceTime = inftime;
189 | }
190 |
191 | void ObjectsRecogFilter::setActiveLabel(QString key, bool value)
192 | {
193 | activeLabels[key] = value;
194 | }
195 |
196 | QMap ObjectsRecogFilter::getActiveLabels()
197 | {
198 | return activeLabels;
199 | }
200 |
201 | bool ObjectsRecogFilter::getActiveLabel(QString key)
202 | {
203 | return activeLabels.value(key,false);
204 | }
205 |
206 | double ObjectsRecogFilter::getAngle() const
207 | {
208 | return ang;
209 | }
210 |
211 | void ObjectsRecogFilter::setAngle(const double value)
212 | {
213 | ang = value;
214 | emit angleChanged();
215 | }
216 |
217 | double ObjectsRecogFilter::getImgHeight()
218 | {
219 | return tf.getHeight();
220 | }
221 |
222 | double ObjectsRecogFilter::getImgWidth()
223 | {
224 | return tf.getWidth();
225 | }
226 |
227 | QImage rotateImage(QImage img, double rotation)
228 | {
229 | QPoint center = img.rect().center();
230 | QMatrix matrix;
231 | matrix.translate(center.x(), center.y());
232 | matrix.rotate(rotation);
233 |
234 | return img.transformed(matrix);
235 | }
236 |
237 | QVideoFrame ObjectsRecogFilterRunable::run(QVideoFrame *input, const QVideoSurfaceFormat &surfaceFormat, RunFlags flags)
238 | {
239 | Q_UNUSED(surfaceFormat);
240 | Q_UNUSED(flags);
241 |
242 | QImage img;
243 | bool mirrorHorizontal;
244 | bool mirrorVertical = false;
245 |
246 | if(input->isValid())
247 | {
248 | // Get image from video frame, we need to convert it
249 | // for unsupported QImage formats, i.e Format_YUV420P
250 | //
251 | // When input has an unsupported format the QImage
252 | // default format is ARGB32
253 | //
254 | // NOTE: BGR images are not properly managed by qt_imageFromVideoFrame
255 | //
256 | bool BGRVideoFrame = AuxUtils::isBGRvideoFrame(*input);
257 | if (BGRVideoFrame)
258 | {
259 | input->map(QAbstractVideoBuffer::ReadOnly);
260 | img = QImage(input->bits(),input->width(),input->height(),QImage::Format_ARGB32).copy();
261 | input->unmap();
262 | // WARNING: Mirror only for Android? How to check if this has to be done?
263 | // surfaceFormat.isMirrored() == false for Android
264 | mirrorVertical = true;
265 | }
266 | else img = qt_imageFromVideoFrame(*input);
267 |
268 | // Check if mirroring is needed
269 | if (!mirrorVertical) mirrorVertical = surfaceFormat.isMirrored();
270 | mirrorHorizontal = surfaceFormat.scanLineDirection() == QVideoSurfaceFormat::BottomToTop;
271 | img = img.mirrored(mirrorHorizontal,mirrorVertical);
272 |
273 | // Check img is valid
274 | if (img.format() != QImage::Format_Invalid)
275 | {
276 | // Take into account the rotation
277 | img = rotateImage(img,-m_filter->getVideoOrientation());
278 |
279 | // Get show inference time
280 | bool showInfTime = m_filter->getShowInfTime();
281 |
282 | QRectF srcRect;
283 |
284 | // Calculate source rect if needed
285 | if (showInfTime)
286 | srcRect = AuxUtils::frameMatchImg(img,m_filter->getContentSize());
287 |
288 | // If not initialized, intialize with image size
289 | if (!m_filter->getInitialized())
290 | m_filter->init(img.height(),img.width());
291 | else if (m_filter->getImgHeight() != img.height() ||
292 | m_filter->getImgWidth() != img.width())
293 | // If image size changed, initialize input tensor
294 | m_filter->initInput(img.height(),img.width());
295 |
296 | // Get a mutex for creating a thread to execute TensorFlow
297 | if (m_filter->getRunning())
298 | {
299 | //img.save("/home/pi/imageTF.png");
300 | emit m_filter->runTensorFlow(img);
301 | }
302 |
303 | // Image segmentation network
304 | if (network == TensorFlow::knIMAGE_SEGMENTATION)
305 | {
306 | // Assumed only 1 masks that must be scaled to the image size
307 | QImage segmentation = masks.first();
308 |
309 | img = AuxUtils::drawSegmentation(img,segmentation);
310 |
311 | // Show inference time
312 | if (showInfTime)
313 | {
314 | QString text = QString::number(inferenceTime) + " ms";
315 | img = AuxUtils::drawText(img,srcRect,text);
316 | }
317 | }
318 | // Object detection network
319 | else if (network == TensorFlow::knOBJECT_DETECTION)
320 | {
321 | // Draw masks on image
322 | if (!masks.isEmpty())
323 | img = AuxUtils::drawMasks(img,img.rect(),results,confidence,boxes,masks,m_filter->getMinConfidence(),m_filter->getActiveLabels());
324 |
325 | // Draw boxes on image
326 | img = AuxUtils::drawBoxes(img,img.rect(),results,confidence,boxes,m_filter->getMinConfidence(),
327 | m_filter->getActiveLabels(),!BGRVideoFrame);
328 |
329 | // Show inference time
330 | if (showInfTime)
331 | {
332 | QString text = QString::number(inferenceTime) + " ms";
333 | img = AuxUtils::drawText(img,srcRect,text);
334 | }
335 |
336 | }
337 |
338 | // Restore rotation
339 | img = rotateImage(img,m_filter->getVideoOrientation());
340 | }
341 |
342 | // NOTE: for BGR images loaded as RGB
343 | if (BGRVideoFrame) img = img.rgbSwapped();
344 |
345 | // Return video frame from img
346 | return QVideoFrame(img);
347 | }
348 |
349 | return *input;
350 | }
351 |
352 | QVideoFilterRunnable *ObjectsRecogFilter::createFilterRunnable()
353 | {
354 | rfr = new ObjectsRecogFilterRunable(this,tf.getResults());
355 | return rfr;
356 | }
357 |
--------------------------------------------------------------------------------
/cpp/auxutils.cpp:
--------------------------------------------------------------------------------
1 | #include "auxutils.h"
2 | #include
3 | #include
4 | #include
5 | #include
6 | #include
7 | #include
8 | #include
9 | #include
10 | #include
11 | #include
12 | #include
13 | #include
14 | #include
15 | #include
16 | #include
17 | #include
18 |
19 | #include "colormanager.h"
20 | #include "math.h"
21 |
22 | double AuxUtils::dpi(QSizeF size)
23 | {
24 | if (QGuiApplication::screens().count() > 0)
25 | {
26 | double dpi = QGuiApplication::screens().first()->logicalDotsPerInch();
27 | double lx = 1*size.width()/QGuiApplication::screens().first()->size().width();
28 | double ly = 1*size.height()/QGuiApplication::screens().first()->size().height();
29 | double ml = 0.5*(lx+ly);
30 |
31 | return dpi * ml;
32 | }
33 | return 160;
34 | }
35 |
36 | // FIXME: properly implement this to be independent of the screen size and resolution
37 | int AuxUtils::sp(int pixel, QSizeF size)
38 | {
39 | //qDebug() << "Physical DPI:" << QApplication::screens().first()->physicalDotsPerInch();
40 | //qDebug() << "Logical DPI:" << QApplication::screens().first()->logicalDotsPerInch();
41 | //qDebug() << "Pixel ratio:" << QApplication::screens().first()->devicePixelRatio();
42 |
43 | // iPhone 7: 1.5
44 | // iPad Mini 4: 1
45 | // Android: 1
46 | // Linux: 4
47 | // iPad Pro: 1
48 | // Raspberry Pi touch screen 7": 1
49 |
50 | return int(pixel * (dpi(size) / 160) * qApp->devicePixelRatio());
51 | }
52 |
53 | QString AuxUtils::deviceInfo()
54 | {
55 | QSysInfo info;
56 |
57 | return info.prettyProductName() + '\n' + '\n' +
58 | QString::number(QThread::idealThreadCount()) + " " + tr("cores");
59 | }
60 |
61 | int AuxUtils::numberThreads()
62 | {
63 | return QThread::idealThreadCount();
64 | }
65 |
66 | QString AuxUtils::qtVersion()
67 | {
68 | return qVersion();
69 | }
70 |
71 | QString AuxUtils::getAssetsPath()
72 | {
73 | return assetsPath;
74 | }
75 |
76 | QImage AuxUtils::drawText(QImage image, QRectF rect, QString text, Qt::AlignmentFlag pos, Qt::GlobalColor borderColor, double borderSize, Qt::GlobalColor fontColor, QFont font)
77 | {
78 | QPainter p;
79 | QRectF r = rect;
80 | QPainterPath path;
81 | QPen pen;
82 | QBrush brush;
83 | QStringList lines;
84 |
85 | if (p.begin(&image))
86 | {
87 | // Configure font
88 | font.setPixelSize(AuxUtils::sp(FONT_PIXEL_SIZE_TEXT,rect.size()));
89 | font.setStyleHint(QFont::Times, QFont::PreferAntialias);
90 |
91 | // Configure pen
92 | pen.setWidthF(borderSize);
93 | pen.setStyle(Qt::SolidLine);
94 | pen.setColor(borderColor);
95 | pen.setCapStyle(Qt::RoundCap);
96 | pen.setJoinStyle(Qt::RoundJoin);
97 |
98 | // Configure brush
99 | brush.setStyle(Qt::SolidPattern);
100 | brush.setColor(fontColor);
101 |
102 | // Get lines
103 | lines = text.split('\n',QString::SkipEmptyParts);
104 |
105 | // Calculate text position
106 | QFontMetrics fm(font);
107 | for(int i=0;i confidences, QList boxes, QList masks,
141 | double minConfidence, QMap activeLabels)
142 | {
143 | Q_UNUSED(activeLabels);
144 | Q_UNUSED(captions);
145 | Q_UNUSED(rect);
146 |
147 | QPainter p;
148 |
149 |
150 | if (p.begin(&image))
151 | {
152 | // http://doc.qt.io/qt-5/qpainter.html#CompositionMode-enum
153 | p.setCompositionMode(QPainter::CompositionMode_SourceOver);
154 |
155 | // Draw each mask
156 | for(int i=0;i= minConfidence && activeLabels[captions[i]])
160 | {
161 | masks[i] = setOpacity(masks[i],MASK_OPACITY);
162 | p.drawImage(boxes[i].topLeft(),masks[i]);
163 | }
164 | }
165 | }
166 | return image;
167 | }
168 |
169 | QImage AuxUtils::drawSegmentation(QImage image, QImage segmentation)
170 | {
171 | QPainter p;
172 |
173 | // Scale segmentation mask to image size
174 | QImage mask = segmentation.scaled(image.size(),Qt::IgnoreAspectRatio,Qt::FastTransformation);
175 |
176 | // Apply mask to image
177 | if (p.begin(&image))
178 | {
179 | // http://doc.qt.io/qt-5/qpainter.html#CompositionMode-enum
180 | p.setCompositionMode(QPainter::CompositionMode_SourceOver);
181 | p.drawImage(0,0,mask);
182 | }
183 |
184 | return image;
185 | }
186 |
187 | QPointF boxCenter(QRectF rect, int offsetX=0, int offsetY=0)
188 | {
189 | return QPointF(rect.left() + rect.width()*0.5 + offsetX, rect.top() + rect.height()*0.5 + offsetY);
190 | }
191 |
192 | QRectF pointCircle(QPointF p, double radius)
193 | {
194 | return QRectF(p.x()-radius,p.y()-radius,2*radius,2*radius);
195 | }
196 |
197 | QRectF pointRect(QPointF p, double width, double height)
198 | {
199 | return QRectF(p.x()-0.5*width,p.y()-0.5*height,width,height);
200 | }
201 |
202 | QPointF midPoint(QPointF a, QPointF b)
203 | {
204 | return QPointF(0.5*(a.x()+b.x()),0.5*(a.y()+b.y()));
205 | }
206 |
207 | bool rectInside(QRectF a, QRectF b)
208 | {
209 | return b.left()>=a.left() && b.top()>=a.top() && b.right()<=a.right() && b.bottom()<=a.bottom();
210 | }
211 |
212 | bool pointInside(QPointF p, QRectF r)
213 | {
214 | return p.x()>=r.left() && p.x()<=r.right() && p.y()<=r.top() && p.y()>=r.bottom();
215 | }
216 |
217 | double getAngle(QPointF a, QPointF b)
218 | {
219 | double angle = atan2(a.y()-b.y(),a.x()-b.x()) * 180 / M_PI;
220 |
221 | // Check it is not: Nan, -Inf or +Inf
222 | angle = angle != angle || angle > std::numeric_limits::max() || angle < -std::numeric_limits::max() ? 0 : angle;
223 |
224 | return angle;
225 | }
226 |
227 | QImage AuxUtils::drawBoxes(QImage image, QRect rect, QStringList captions, QList confidences, QList boxes, double minConfidence,
228 | QMap activeLabels, bool rgb)
229 | {
230 | Q_UNUSED(rect);
231 |
232 | ColorManager cm;
233 | QPainter p;
234 | QBrush brush;
235 | QPen pen;
236 | QFont font;
237 | QPen fPen;
238 | QBrush bBrush;
239 | QPen bPen;
240 |
241 | if (p.begin(&image))
242 | {
243 | // Configure pen
244 | pen.setStyle(Qt::SolidLine);
245 | pen.setWidthF(LINE_WIDTH);
246 |
247 | // Configure font pen
248 | fPen.setStyle(Qt::SolidLine);
249 | fPen.setColor(Qt::black);
250 |
251 | // Configure back pen
252 | bPen.setStyle(Qt::SolidLine);
253 |
254 | // Configure brush
255 | brush.setStyle(Qt::NoBrush);
256 |
257 | // Configure back brush
258 | bBrush.setStyle(Qt::SolidPattern);
259 |
260 | // Configure font
261 | font.setCapitalization(QFont::Capitalize);
262 | font.setPixelSize(AuxUtils::sp(FONT_PIXEL_SIZE_BOX,rect.size()));
263 |
264 | // Configure painter
265 | p.setRenderHint(QPainter::Antialiasing);
266 | p.setFont(font);
267 |
268 | QFontMetrics fm(font);
269 |
270 | // Draw each box
271 | for(int i=0;i= minConfidence && activeLabels[captions[i]])
275 | {
276 | // Draw box
277 | cm.setRgb(rgb);
278 | pen.setColor(cm.getColor(captions[i]));
279 | p.setPen(pen);
280 | p.setBrush(brush);
281 | p.drawRect(boxes[i]);
282 |
283 | // Format text
284 | QString confVal = QString::number(qRound(confidences[i] * 100)) + " %";
285 | QString text = captions[i] + " - " + confVal;
286 |
287 | // Text rect
288 | int width = fm.width(text)+FONT_WIDTH_MARGIN;
289 | int height = fm.height();
290 | int left = boxes[i].left()>=0 ? int(boxes[i].left()) : int(boxes[i].right()-width);
291 | int top = boxes[i].top()-fm.height()>=0 ? int(boxes[i].top()-fm.height()) : int(boxes[i].bottom());
292 |
293 | // Text position
294 | int tLeft = left+FONT_WIDTH_MARGIN/2;
295 | int tTop = boxes[i].top()-fm.height()>=0 ? int(boxes[i].top() - FONT_HEIGHT_MARGIN) : int(boxes[i].bottom() + height - FONT_HEIGHT_MARGIN);
296 |
297 | // Draw text background
298 | bPen.setColor(pen.color());
299 | bBrush.setColor(pen.color());
300 | p.setPen(bPen);
301 | p.setBrush(bBrush);
302 | p.drawRect(left,top,width,height);
303 |
304 | // Draw tex
305 | p.setPen(fPen);
306 | p.drawText(tLeft,tTop,text);
307 | }
308 | }
309 | }
310 |
311 | return image;
312 | }
313 |
314 | QString AuxUtils::getDefaultModelFilename()
315 | {
316 | return assetsPath + QDir::separator() + modelName;
317 | }
318 |
319 | QString AuxUtils::getDefaultLabelsFilename()
320 | {
321 | return assetsPath + QDir::separator() + labelsName;
322 | }
323 |
324 | QRectF AuxUtils::frameMatchImg(QImage img, QSize rectSize)
325 | {
326 | QSize isize = img.size();
327 | rectSize.scale(isize, Qt::KeepAspectRatio);
328 | QPoint center = img.rect().center();
329 |
330 | return QRectF(center.x()-rectSize.width()*0.5,center.y()-rectSize.height()*0.5,rectSize.width(),rectSize.height());
331 | }
332 |
333 | bool AuxUtils::readLabels(QString filename)
334 | {
335 | if (!filename.trimmed().isEmpty())
336 | {
337 | QFile textFile(filename);
338 |
339 | if (textFile.exists())
340 | {
341 | QByteArray line;
342 |
343 | textFile.open(QIODevice::ReadOnly);
344 |
345 | line = textFile.readLine().trimmed();
346 | while(!line.isEmpty()) // !textFile.atEnd() &&
347 | {
348 | labels.append(line);
349 | line = textFile.readLine().trimmed();
350 | }
351 |
352 | textFile.close();
353 | if (labels.count()>0) labels.removeFirst();
354 | return true;
355 | }
356 | }
357 | return false;
358 | }
359 |
360 | QStringList AuxUtils::getLabels()
361 | {
362 | if (labels.isEmpty()) readLabels(AuxUtils::getDefaultLabelsFilename());
363 |
364 | return labels;
365 | }
366 |
367 | bool AuxUtils::isBGRvideoFrame(QVideoFrame f)
368 | {
369 | return f.pixelFormat() == QVideoFrame::Format_BGRA32 ||
370 | f.pixelFormat() == QVideoFrame::Format_BGRA32_Premultiplied ||
371 | f.pixelFormat() == QVideoFrame::Format_BGR32 ||
372 | f.pixelFormat() == QVideoFrame::Format_BGR24 ||
373 | f.pixelFormat() == QVideoFrame::Format_BGR565 ||
374 | f.pixelFormat() == QVideoFrame::Format_BGR555 ||
375 | f.pixelFormat() == QVideoFrame::Format_BGRA5658_Premultiplied;
376 | }
377 |
378 | bool AuxUtils::isBGRimage(QImage i)
379 | {
380 | return i.format() == QImage::Format_BGR30 ||
381 | i.format() == QImage::Format_A2BGR30_Premultiplied;
382 | }
383 |
384 | QVariantList AuxUtils::networkInterfaces()
385 | {
386 | QVariantList list;
387 |
388 | foreach(QNetworkInterface interface, QNetworkInterface::allInterfaces())
389 | {
390 | if (!(interface.flags() & QNetworkInterface::IsLoopBack))
391 | {
392 | QString info;
393 |
394 | info = interface.humanReadableName() + " (" + interface.name() + ") - " + interface.hardwareAddress() + " - " +
395 | (interface.addressEntries().count()>0 ? interface.addressEntries().first().ip().toString() : "None");
396 |
397 | list << info;
398 | }
399 | }
400 |
401 | return list;
402 | }
403 |
404 | void AuxUtils::setAngleHor(double angle) { AuxUtils::angleHor = angle;}
405 | void AuxUtils::setAngleVer(double angle) { AuxUtils::angleVer = angle;}
406 |
407 | bool AuxUtils::setResolution(QString res)
408 | {
409 | QStringList sRes = res.split(RES_CHAR);
410 |
411 | if (sRes.count()>1)
412 | {
413 | width = sRes[0].toInt();
414 | height = sRes[1].toInt();
415 | }
416 | return false;
417 | }
418 |
--------------------------------------------------------------------------------
/Configuration.qml:
--------------------------------------------------------------------------------
1 | import QtQuick 2.9
2 | import QtQuick.Controls 2.2
3 |
4 | Page {
5 | id: root
6 | title: qsTr("Settings")
7 |
8 | property double minConfidence
9 | property int nThreads
10 | property bool showInfTime
11 | property string resolution
12 | property var resolutions: []
13 | property bool loadingRes
14 | property bool semiTransparent
15 | property bool showBackground
16 |
17 | readonly property int leftMargin: 100
18 | readonly property int rightMargin: 100
19 |
20 | signal resolutionUpdated()
21 |
22 | onResolutionsChanged: {
23 | console.log("Camera resolutions: " + resolutions)
24 | loadingRes = true
25 | sbResolution.to = resolutions.length - 1
26 | sbResolution.items = resolutions
27 | sbResolution.value = -1
28 | loadingRes = false
29 | sbResolution.value = sbResolution.valueFromText(resolution)
30 | }
31 |
32 | // TabBar
33 | footer: TabBar {
34 | id: tabBar
35 | currentIndex: swipeView.currentIndex
36 |
37 | TabButton {
38 | text: qsTr("Neural Network && Camera")
39 | }
40 |
41 | TabButton {
42 | text: qsTr("Screen info")
43 | }
44 |
45 | TabButton {
46 | text: qsTr("Hardware info && Close app")
47 | }
48 | }
49 |
50 |
51 |
52 | // tab contents
53 | SwipeView {
54 | id: swipeView
55 | anchors.fill: parent
56 | clip: true
57 | currentIndex: tabBar.currentIndex
58 |
59 | Flickable {
60 | contentHeight: column.height
61 | contentWidth: column.width
62 |
63 | flickableDirection: Flickable.VerticalFlick
64 |
65 | Column{
66 | id: column
67 | width: root.width
68 |
69 | Item{
70 | height: 20
71 | width: 1
72 | }
73 |
74 | Text{
75 | anchors.leftMargin: leftMargin
76 | anchors.rightMargin: rightMargin
77 | anchors.horizontalCenter: parent.horizontalCenter
78 | horizontalAlignment: Text.AlignHCenter
79 | width: parent.width
80 | wrapMode: Text.WordWrap
81 | elide: Text.ElideRight
82 | text: qsTr("Minimum confidence")
83 | }
84 |
85 | Item{
86 | width: 1
87 | height: 10
88 | }
89 |
90 | Slider{
91 | id: slider
92 | anchors.horizontalCenter: parent.horizontalCenter
93 | width: parent.width - (leftMargin+rightMargin)
94 | from: 0
95 | to: 1
96 | value: minConfidence
97 | live: true
98 |
99 | onValueChanged: minConfidence = value
100 | }
101 |
102 | Text {
103 | anchors.leftMargin: leftMargin
104 | anchors.rightMargin: rightMargin
105 | anchors.horizontalCenter: parent.horizontalCenter
106 | horizontalAlignment: Text.AlignHCenter
107 | width: parent.width
108 | wrapMode: Text.WordWrap
109 | elide: Text.ElideRight
110 | text: Math.round(slider.position * 100) + " %"
111 | }
112 |
113 | Item{
114 | width: 1
115 | height: 10
116 | }
117 |
118 | Row {
119 | width: parent.width
120 | spacing: 2
121 |
122 | Item{
123 | height: 1
124 | width: (parent.width - parent.spacing - tThread.width)*0.5
125 | }
126 |
127 | Text{
128 | id: tThread
129 | anchors.leftMargin: 30
130 | anchors.verticalCenter: parent.verticalCenter
131 | verticalAlignment: Text.AlignVCenter
132 | wrapMode: Text.WordWrap
133 | elide: Text.ElideRight
134 | text: qsTr("Number of threads")
135 | }
136 | }
137 |
138 | Slider{
139 | id: sThreads
140 | anchors.horizontalCenter: parent.horizontalCenter
141 | width: parent.width - (leftMargin+rightMargin)
142 | from: 1
143 | to: auxUtils.numberThreads()
144 | enabled: to>1
145 | live: true
146 | snapMode: Slider.SnapAlways
147 | stepSize: 1
148 | value: nThreads
149 |
150 | onValueChanged: nThreads = value
151 | }
152 |
153 | Text {
154 | anchors.leftMargin: leftMargin
155 | anchors.rightMargin: rightMargin
156 | anchors.horizontalCenter: parent.horizontalCenter
157 | horizontalAlignment: Text.AlignHCenter
158 | width: parent.width
159 | wrapMode: Text.WordWrap
160 | elide: Text.ElideRight
161 | text: sThreads.value + " " + (sThreads.value>1 ? qsTr("threads") : qsTr("thread"))
162 | }
163 |
164 | Item{
165 | width: 1
166 | height: 20
167 | }
168 |
169 | Text{
170 | anchors.leftMargin: leftMargin
171 | anchors.rightMargin: rightMargin
172 | anchors.horizontalCenter: parent.horizontalCenter
173 | horizontalAlignment: Text.AlignHCenter
174 | width: parent.width
175 | wrapMode: Text.WordWrap
176 | elide: Text.ElideRight
177 | text: qsTr("Camera resolution")
178 | }
179 |
180 | Item{
181 | width: 1
182 | height: 10
183 | }
184 |
185 | SpinBox {
186 | id: sbResolution
187 | anchors.horizontalCenter: parent.horizontalCenter
188 | from: 0
189 | editable: false
190 | property var items
191 |
192 | textFromValue: function(value) {
193 | return items[value];
194 | }
195 |
196 | valueFromText: function(text) {
197 | for (var i = 0; i < items.length; i++)
198 | {
199 | if (items[i].toLowerCase() === text.toLowerCase())
200 | return i
201 | }
202 | return value
203 | }
204 |
205 | onValueChanged:
206 | if (value>=0 && !loadingRes)
207 | {
208 | resolution = textFromValue(value)
209 | console.log("Resolution to save: " + resolution)
210 | resolutionUpdated()
211 | }
212 | }
213 |
214 | Item{
215 | height: 30
216 | width: 1
217 | }
218 | }
219 | }
220 |
221 | Flickable {
222 | id: flickInfo
223 | contentHeight: column3.height
224 | contentWidth: column3.width
225 |
226 | flickableDirection: Flickable.VerticalFlick
227 |
228 | Column{
229 | id: column3
230 | width: root.width
231 |
232 | Item{
233 | height: 20
234 | width: 1
235 | }
236 |
237 | Text{
238 | anchors.leftMargin: leftMargin
239 | anchors.rightMargin: rightMargin
240 | anchors.horizontalCenter: parent.horizontalCenter
241 | horizontalAlignment: Text.AlignHCenter
242 | width: parent.width
243 | wrapMode: Text.WordWrap
244 | elide: Text.ElideRight
245 | text: qsTr("General information")
246 | }
247 |
248 | Item{
249 | width: 1
250 | height: 10
251 | }
252 |
253 | Row{
254 | width: parent.width - (leftMargin+rightMargin)
255 | anchors.horizontalCenter: parent.horizontalCenter
256 | spacing: width - tShowInfTime.width - sShowInfTime.width
257 |
258 | Text {
259 | id: tShowInfTime
260 | text: qsTr("Show inference time")
261 | anchors.verticalCenter: parent.verticalCenter
262 | verticalAlignment: Text.AlignVCenter
263 | }
264 |
265 | Switch{
266 | anchors.verticalCenter: parent.verticalCenter
267 | id: sShowInfTime
268 | checked: showInfTime
269 |
270 | onToggled: showInfTime = checked
271 | }
272 | }
273 |
274 | Item{
275 | height: 20
276 | width: 1
277 | }
278 |
279 | Text{
280 | anchors.leftMargin: leftMargin
281 | anchors.rightMargin: rightMargin
282 | anchors.horizontalCenter: parent.horizontalCenter
283 | horizontalAlignment: Text.AlignHCenter
284 | width: parent.width
285 | wrapMode: Text.WordWrap
286 | elide: Text.ElideRight
287 | text: qsTr("Segmentation")
288 | }
289 |
290 | Item{
291 | width: 1
292 | height: 10
293 | }
294 |
295 | CheckBox{
296 | id: chbSemiTransparent
297 | anchors.horizontalCenter: parent.horizontalCenter
298 | text: "Semi-transparent objects"
299 | checked: semiTransparent
300 | onCheckStateChanged: semiTransparent = checked
301 | }
302 |
303 | CheckBox{
304 | anchors.left: chbSemiTransparent.left
305 | text: "Show real background"
306 | checked: showBackground
307 | onCheckStateChanged: showBackground = checked
308 | }
309 | }
310 | }
311 |
312 | Flickable {
313 | contentHeight: column2.height
314 | contentWidth: column2.width
315 |
316 | flickableDirection: Flickable.VerticalFlick
317 |
318 | Column{
319 | id: column2
320 | width: root.width
321 |
322 | Item{
323 | height: 20
324 | width: 1
325 | }
326 |
327 | Text{
328 | anchors.leftMargin: leftMargin
329 | anchors.rightMargin: rightMargin
330 | anchors.horizontalCenter: parent.horizontalCenter
331 | horizontalAlignment: Text.AlignHCenter
332 | width: parent.width
333 | wrapMode: Text.WordWrap
334 | elide: Text.ElideRight
335 | height: 30
336 | text: qsTr("Hardware info")
337 | }
338 |
339 | Item{
340 | width: 1
341 | height: 10
342 | }
343 |
344 | Column{
345 | width: root.width
346 |
347 | Repeater{
348 | model: auxUtils.networkInterfaces()
349 |
350 | Text{
351 | anchors.leftMargin: leftMargin
352 | anchors.rightMargin: rightMargin
353 | horizontalAlignment: Text.AlignHCenter
354 | width: parent.width
355 | height: 30
356 | wrapMode: Text.WordWrap
357 | elide: Text.ElideRight
358 | text: modelData
359 | }
360 | }
361 | }
362 |
363 | Item{
364 | width: 1
365 | height: 30
366 | }
367 |
368 | Button{
369 | id: bClose
370 | anchors.horizontalCenter: parent.horizontalCenter
371 | text: qsTr("Close app")
372 | onClicked: dialog.open()
373 |
374 | contentItem: Text {
375 | text: bClose.text
376 | font: bClose.font
377 | opacity: enabled ? 1.0 : 0.3
378 | color: "white"
379 | horizontalAlignment: Text.AlignHCenter
380 | verticalAlignment: Text.AlignVCenter
381 | elide: Text.ElideRight
382 | }
383 |
384 | background: Rectangle {
385 | implicitWidth: 100
386 | implicitHeight: 40
387 | color: bClose.down ? "#ff0000" : "#aa0000"
388 | border.color: "#7e181a"
389 | border.width: 1
390 | radius: 0
391 | }
392 | }
393 | }
394 | }
395 |
396 | }
397 |
398 | Rectangle{
399 | id: backPageIndicator
400 | anchors.bottom: parent.bottom
401 | height: 20
402 | width: parent.width
403 | }
404 |
405 | PageIndicator {
406 | id: pageIndicator
407 | count: swipeView.count
408 | currentIndex: swipeView.currentIndex
409 | anchors.centerIn: backPageIndicator
410 |
411 | delegate: Rectangle{
412 | implicitWidth: 10
413 | implicitHeight: 10
414 | radius: width
415 | }
416 | }
417 |
418 | Dialog {
419 | id: dialog
420 | x: 0.5*(parent.width - width)
421 | y: 0.5*(parent.height - height)
422 | title: "Close app"
423 | standardButtons: Dialog.Ok | Dialog.Cancel
424 | modal: true
425 |
426 | Label{
427 | text: qsTr("Do you really want to close this app?")
428 | }
429 |
430 | onAccepted: Qt.callLater(Qt.quit)
431 | onRejected: dialog.close()
432 | }
433 | }
434 |
--------------------------------------------------------------------------------
/cpp/tensorflow.cpp:
--------------------------------------------------------------------------------
1 | #include "tensorflow.h"
2 |
3 | #include "tensorflow/lite/kernels/internal/tensor.h"
4 | #include "tensorflow/lite/kernels/internal/tensor_utils.h"
5 |
6 | #include "get_top_n.h"
7 | #include "colormanager.h"
8 |
9 | #include
10 | #include
11 | #include
12 | #include
13 | #include
14 | #include
15 | #include
16 |
17 | TensorFlow::TensorFlow(QObject *parent) : QObject(parent)
18 | {
19 | initialized = false;
20 | accelaration = false;
21 | verbose = true;
22 | numThreads = 1;
23 | threshold = 0.1;
24 | has_detection_masks = false;
25 | }
26 | TensorFlow::~TensorFlow()
27 | {}
28 |
29 | template
30 | bool formatImageQt(T* out, QImage image, int image_channels, int wanted_height, int wanted_width, int wanted_channels, bool input_floating, bool scale = false)
31 | {
32 | const float input_mean = 127.5f;
33 | const float input_std = 127.5f;
34 |
35 | // Check same number of channels
36 | if (image_channels != wanted_channels)
37 | {
38 | qDebug() << "ERROR: the image has" << image_channels << " channels. Wanted channels:" << wanted_channels;
39 | return false;
40 | }
41 |
42 | // Scale image if needed
43 | if (scale && (image.width() != wanted_width || image.height() != wanted_height))
44 | image = image.scaled(wanted_height,wanted_width,Qt::IgnoreAspectRatio,Qt::FastTransformation);
45 |
46 | // Number of pixels
47 | const int numberPixels = image.height()*image.width()*wanted_channels;
48 |
49 | // Pointer to image data
50 | const uint8_t *output = image.bits();
51 |
52 | // Boolean to [0,1]
53 | const int inputFloat = input_floating ? 1 : 0;
54 | const int inputInt = input_floating ? 0 : 1;
55 |
56 | // Transform to [0,128] ¿?
57 | for (int i = 0; i < numberPixels; i++)
58 | {
59 | out[i] = inputFloat*((output[i] - input_mean) / input_std) + // inputFloat*(output[i]/ 128.f - 1.f) +
60 | inputInt*(uint8_t)output[i];
61 | //qDebug() << out[i];
62 | }
63 |
64 | return true;
65 | }
66 |
67 | // -----------------------------------------------------------------------------------------------------------------------
68 | // https://github.com/tensorflow/tensorflow/blob/master/tensorflow/contrib/lite/examples/label_image/bitmap_helpers_impl.h
69 | // -----------------------------------------------------------------------------------------------------------------------
70 | template
71 | void formatImageTFLite(T* out, const uint8_t* in, int image_height, int image_width, int image_channels, int wanted_height, int wanted_width, int wanted_channels, bool input_floating)
72 | {
73 | const float input_mean = 127.5f;
74 | const float input_std = 127.5f;
75 |
76 | int number_of_pixels = image_height * image_width * image_channels;
77 | std::unique_ptr interpreter(new Interpreter);
78 |
79 | int base_index = 0;
80 |
81 | // two inputs: input and new_sizes
82 | interpreter->AddTensors(2, &base_index);
83 |
84 | // one output
85 | interpreter->AddTensors(1, &base_index);
86 |
87 | // set input and output tensors
88 | interpreter->SetInputs({0, 1});
89 | interpreter->SetOutputs({2});
90 |
91 | // set parameters of tensors
92 | TfLiteQuantizationParams quant;
93 | interpreter->SetTensorParametersReadWrite(0, kTfLiteFloat32, "input", {1, image_height, image_width, image_channels}, quant);
94 | interpreter->SetTensorParametersReadWrite(1, kTfLiteInt32, "new_size", {2},quant);
95 | interpreter->SetTensorParametersReadWrite(2, kTfLiteFloat32, "output", {1, wanted_height, wanted_width, wanted_channels}, quant);
96 |
97 | ops::builtin::BuiltinOpResolver resolver;
98 | const TfLiteRegistration *resize_op = resolver.FindOp(BuiltinOperator_RESIZE_BILINEAR,1);
99 | auto* params = reinterpret_cast(malloc(sizeof(TfLiteResizeBilinearParams)));
100 | params->align_corners = false;
101 | interpreter->AddNodeWithParameters({0, 1}, {2}, nullptr, 0, params, resize_op, nullptr);
102 | interpreter->AllocateTensors();
103 |
104 |
105 | // fill input image
106 | // in[] are integers, cannot do memcpy() directly
107 | auto input = interpreter->typed_tensor(0);
108 | for (int i = 0; i < number_of_pixels; i++)
109 | input[i] = in[i];
110 |
111 | // fill new_sizes
112 | interpreter->typed_tensor(1)[0] = wanted_height;
113 | interpreter->typed_tensor(1)[1] = wanted_width;
114 |
115 | interpreter->Invoke();
116 |
117 | auto output = interpreter->typed_tensor(2);
118 | auto output_number_of_pixels = wanted_height * wanted_height * wanted_channels;
119 |
120 | for (int i = 0; i < output_number_of_pixels; i++)
121 | {
122 | if (input_floating)
123 | out[i] = (output[i] - input_mean) / input_std;
124 | else
125 | out[i] = (uint8_t)output[i];
126 | }
127 | }
128 | bool TensorFlow::init(int imgHeight, int imgWidth)
129 | {
130 | if (!initialized)
131 | initialized = initTFLite(imgHeight,imgWidth);
132 |
133 | return initialized;
134 | }
135 |
136 | void TensorFlow::initInput(int imgHeight, int imgWidth)
137 | {
138 | Q_UNUSED(imgHeight);
139 | Q_UNUSED(imgWidth);
140 | }
141 |
142 | // ------------------------------------------------------------------------------------------------------------------------------
143 | // Adapted from: https://github.com/tensorflow/tensorflow/blob/master/tensorflow/contrib/lite/examples/label_image/label_image.cc
144 | // ------------------------------------------------------------------------------------------------------------------------------
145 | bool TensorFlow::initTFLite(int imgHeight, int imgWidth)
146 | {
147 | Q_UNUSED(imgHeight);
148 | Q_UNUSED(imgWidth);
149 |
150 | try{
151 | // Open model & assign error reporter
152 | model = AuxUtils::getDefaultModelFilename().trimmed().isEmpty() && AuxUtils::getDefaultLabelsFilename().trimmed().isEmpty() ? nullptr :
153 | FlatBufferModel::BuildFromFile(filename.toStdString().c_str(),&error_reporter);
154 |
155 | if(model == nullptr)
156 | {
157 | qDebug() << "TensorFlow model loading: ERROR";
158 | return false;
159 | }
160 |
161 | // Link model & resolver
162 | InterpreterBuilder builder(*model.get(), resolver);
163 |
164 | // Check interpreter
165 | if(builder(&interpreter) != kTfLiteOk)
166 | {
167 | qDebug() << "Interpreter: ERROR";
168 | return false;
169 | }
170 |
171 | // Apply accelaration (Neural Network Android)
172 | interpreter->UseNNAPI(accelaration);
173 |
174 | if(interpreter->AllocateTensors() != kTfLiteOk)
175 | {
176 | qDebug() << "Allocate tensors: ERROR";
177 | return false;
178 | }
179 |
180 | // Set kind of network
181 | // WARNING: It is assumed the kind of network depending on the number of outputs
182 | kind_network = interpreter->outputs().size()>1 ? knOBJECT_DETECTION : knIMAGE_SEGMENTATION;
183 |
184 | if (verbose)
185 | {
186 | int i_size = interpreter->inputs().size();
187 | int o_size = interpreter->outputs().size();
188 | int t_size = interpreter->tensors_size();
189 |
190 | qDebug() << "tensors size: " << t_size;
191 | qDebug() << "nodes size: " << interpreter->nodes_size();
192 | qDebug() << "inputs: " << i_size;
193 | qDebug() << "outputs: " << o_size;
194 |
195 | for (int i = 0; i < i_size; i++)
196 | qDebug() << "input" << i << "name:" << interpreter->GetInputName(i) << ", type:" << interpreter->tensor(interpreter->inputs()[i])->type;
197 |
198 | for (int i = 0; i < o_size; i++)
199 | qDebug() << "output" << i << "name:" << interpreter->GetOutputName(i) << ", type:" << interpreter->tensor(interpreter->outputs()[i])->type;
200 |
201 | // for (int i = 0; i < t_size; i++)
202 | // {
203 | // if (interpreter->tensor(i)->name)
204 | // qDebug() << i << ":" << interpreter->tensor(i)->name << ","
205 | // << interpreter->tensor(i)->bytes << ","
206 | // << interpreter->tensor(i)->type << ","
207 | // << interpreter->tensor(i)->params.scale << ","
208 | // << interpreter->tensor(i)->params.zero_point;
209 | // }
210 | }
211 |
212 | // Get input dimension from the input tensor metadata
213 | // Assuming one input only
214 | int input = interpreter->inputs()[0];
215 | TfLiteIntArray* dims = interpreter->tensor(input)->dims;
216 |
217 | // Save outputs
218 | outputs.clear();
219 | for(unsigned int i=0;ioutputs().size();i++)
220 | outputs.push_back(interpreter->tensor(interpreter->outputs()[i]));
221 |
222 | wanted_height = dims->data[1];
223 | wanted_width = dims->data[2];
224 | wanted_channels = dims->data[3];
225 |
226 | if (verbose)
227 | {
228 | qDebug() << "Wanted height:" << wanted_height;
229 | qDebug() << "Wanted width:" << wanted_width;
230 | qDebug() << "Wanted channels:" << wanted_channels;
231 | }
232 |
233 | if (numThreads > 1)
234 | interpreter->SetNumThreads(numThreads);
235 |
236 | // Read labels
237 | if (readLabels()) qDebug() << "There are" << labels.count() << "labels.";
238 | else qDebug() << "There are NO labels";
239 |
240 | qDebug() << "Tensorflow initialization: OK";
241 | return true;
242 |
243 | }catch(...)
244 | {
245 | qDebug() << "Exception loading model";
246 | return false;
247 | }
248 | }
249 |
250 | // --------------------------------------------------------------------------------------
251 | // Code from: https://github.com/YijinLiu/tf-cpu/blob/master/benchmark/obj_detect_lite.cc
252 | // --------------------------------------------------------------------------------------
253 | template
254 | T* TensorData(TfLiteTensor* tensor, int batch_index);
255 |
256 | template<>
257 | float* TensorData(TfLiteTensor* tensor, int batch_index) {
258 | int nelems = 1;
259 | for (int i = 1; i < tensor->dims->size; i++) nelems *= tensor->dims->data[i];
260 | switch (tensor->type) {
261 | case kTfLiteFloat32:
262 | return tensor->data.f + nelems * batch_index;
263 | default:
264 | qDebug() << "Should not reach here!";
265 | }
266 | return nullptr;
267 | }
268 |
269 | template<>
270 | uint8_t* TensorData(TfLiteTensor* tensor, int batch_index) {
271 | int nelems = 0;
272 | for (int i = 1; i < tensor->dims->size; i++) nelems *= tensor->dims->data[i];
273 | switch (tensor->type) {
274 | case kTfLiteUInt8:
275 | return tensor->data.uint8 + nelems * batch_index;
276 | default:
277 | qDebug() << "Should not reach here!";
278 | }
279 | return nullptr;
280 | }
281 |
282 | int TensorFlow::getKindNetwork()
283 | {
284 | return kind_network;
285 | }
286 |
287 | double TensorFlow::getThreshold() const
288 | {
289 | return threshold;
290 | }
291 |
292 | void TensorFlow::setThreshold(double value)
293 | {
294 | threshold = value;
295 | }
296 |
297 | bool TensorFlow::setInputs(QImage image)
298 | {
299 | return setInputsTFLite(image);
300 | }
301 |
302 | bool TensorFlow::setInputsTFLite(QImage image)
303 | {
304 | // Get inputs
305 | std::vector inputs = interpreter->inputs();
306 |
307 | // Set inputs
308 | for(unsigned int i=0;iinputs().size();i++)
309 | {
310 | int input = inputs[i];
311 |
312 | // Convert input
313 | switch (interpreter->tensor(input)->type)
314 | {
315 | case kTfLiteFloat32:
316 | {
317 | formatImageTFLite(interpreter->typed_tensor(input),image.bits(), image.height(),
318 | image.width(), img_channels, wanted_height, wanted_width,wanted_channels, true);
319 | //formatImageQt(interpreter->typed_tensor(input),image,img_channels,
320 | // wanted_height,wanted_width,wanted_channels,true,true);
321 | break;
322 | }
323 | case kTfLiteUInt8:
324 | {
325 | formatImageTFLite(interpreter->typed_tensor(input),image.bits(),
326 | img_height, img_width, img_channels, wanted_height,
327 | wanted_width, wanted_channels, false);
328 |
329 | //formatImageQt(interpreter->typed_tensor(input),image,img_channels,
330 | // wanted_height,wanted_width,wanted_channels,false);
331 | break;
332 | }
333 | default:
334 | {
335 | qDebug() << "Cannot handle input type" << interpreter->tensor(input)->type << "yet";
336 | return false;
337 | }
338 | }
339 | }
340 |
341 | return true;
342 | }
343 |
344 | bool TensorFlow::inference()
345 | {
346 | return inferenceTFLite();
347 | }
348 |
349 | bool TensorFlow::inferenceTFLite()
350 | {
351 | // Invoke interpreter
352 | if (interpreter->Invoke() != kTfLiteOk)
353 | {
354 | qDebug() << "Failed to invoke interpreter";
355 | return false;
356 | }
357 | return true;
358 | }
359 |
360 | bool TensorFlow::getSegmentationOutputs(QImage &segmentation)
361 | {
362 | const int nObjects = 21;
363 |
364 | // Check one output
365 | if (outputs.size()==1)
366 | {
367 | const float *data = TensorData(outputs.front(), 0);
368 |
369 | segmentation.fill(getShowBackground() ? Qt::transparent : background_color);
370 | for(int i=0;imax) {
381 | object_index = k;
382 | max = value;
383 | }
384 | }
385 | QColor color = objects_colors[object_index];
386 | if (color != background_color)
387 | {
388 | color.setAlphaF(getSemiTransparent() ? object_alpha : 1);
389 | segmentation.setPixelColor(j,i,color);
390 | }
391 | }
392 | }
393 | return true;
394 | }
395 | return false;
396 | }
397 |
398 | bool TensorFlow::getObjectOutputs(QStringList &captions, QList &confidences, QList &locations, QList &masks)
399 | {
400 | if (outputs.size() >= 4)
401 | {
402 | const int num_detections = *TensorData(outputs[3], 0);
403 | const float* detection_classes = TensorData(outputs[1], 0);
404 | const float* detection_scores = TensorData(outputs[2], 0);
405 | const float* detection_boxes = TensorData(outputs[0], 0);
406 | const float* detection_masks = !has_detection_masks || outputs.size()<5 ? nullptr : TensorData(outputs[4], 0);
407 | ColorManager cm;
408 |
409 | for (int i=0; idims->data[2];
441 | const int dim2 = outputs[4]->dims->data[3];
442 | QImage mask(dim1,dim2,QImage::Format_ARGB32_Premultiplied);
443 |
444 | // Set binary mask [dim1,dim2]
445 | for(int j=0;j= MASK_THRESHOLD ?
448 | cm.getColor(label).rgba() : QColor(Qt::transparent).rgba());
449 | // Billinear interpolation
450 | // https://chu24688.tian.yam.com/posts/44797337
451 | //QImage maskScaled = ColorManager::billinearInterpolation(mask,box.height(),box.width());
452 |
453 | // Scale mask to box size
454 | QImage maskScaled = mask.scaled(box.width(),box.height(),Qt::IgnoreAspectRatio,Qt::FastTransformation);
455 |
456 | // Border detection
457 | //QTransform trans(-1,0,1,-2,0,2,-1,0,1);
458 | //maskScaled = ColorManager::applyTransformation(maskScaled,trans);
459 |
460 | // Append to masks
461 | masks.append(maskScaled);
462 | }
463 |
464 | // Save remaining data
465 | captions.append(label);
466 | confidences.append(score);
467 | locations.append(box);
468 | }
469 |
470 | return true;
471 | }
472 | return false;
473 | }
474 |
475 | bool TensorFlow::getSemiTransparent() const
476 | {
477 | return semiTransparent;
478 | }
479 |
480 | void TensorFlow::setSemiTransparent(bool value)
481 | {
482 | semiTransparent = value;
483 | }
484 |
485 | bool TensorFlow::getShowBackground() const
486 | {
487 | return showBackground;
488 | }
489 |
490 | void TensorFlow::setShowBackground(bool value)
491 | {
492 | showBackground = value;
493 | }
494 |
495 | // ---------------------------------------------------------------------------------------------------------------
496 | // Adapted from: https://github.com/tensorflow/tensorflow/tree/master/tensorflow/contrib/lite/examples/label_image
497 | // ---------------------------------------------------------------------------------------------------------------
498 | bool TensorFlow::run(QImage img)
499 | {
500 | QElapsedTimer timer;
501 |
502 | if (initialized)
503 | {
504 | // Start timer
505 | //timer.start();
506 |
507 | // Transform image format & copy data
508 | QImage image = img.format() == format ? img : img.convertToFormat(format);
509 |
510 | // Store original image properties
511 | img_width = image.width();
512 | img_height = image.height();
513 | img_channels = numChannels;
514 |
515 | // Set inputs
516 | if (!setInputs(image)) return false;
517 |
518 | // Perform inference
519 | timer.start();
520 | if (!inference()) return false;
521 | inferenceTime = timer.elapsed();
522 |
523 | // -------------------------------------
524 | // Outputs depend on the kind of network
525 | // -------------------------------------
526 | rCaption.clear();
527 | rConfidence.clear();
528 | rBox.clear();
529 | rMasks.clear();
530 |
531 | // Image classifier
532 | if (kind_network == knIMAGE_SEGMENTATION)
533 | {
534 | QImage segmentation(wanted_height,wanted_width,QImage::Format_ARGB32);
535 |
536 | if (!getSegmentationOutputs(segmentation)) return false;
537 | rMasks.append(segmentation);
538 | }
539 | // Object detection
540 | else if (kind_network == knOBJECT_DETECTION)
541 | {
542 | if (!getObjectOutputs(rCaption,rConfidence,rBox,rMasks)) return false;
543 | }
544 |
545 | //inferenceTime = timer.elapsed();
546 | if (verbose) qDebug() << "Elapsed time: " << inferenceTime << "milliseconds";
547 |
548 | return true;
549 | }
550 |
551 | return false;
552 | }
553 |
554 | // WARNING: function repeated in AuxUtils
555 | bool TensorFlow::readLabels()
556 | {
557 | if (!labelsFilename.trimmed().isEmpty())
558 | {
559 | QFile textFile(labelsFilename);
560 |
561 | if (textFile.exists())
562 | {
563 | QByteArray line;
564 |
565 | labels.clear();
566 | textFile.open(QIODevice::ReadOnly);
567 |
568 | line = textFile.readLine().trimmed();
569 | while(!line.isEmpty()) // !textFile.atEnd() &&
570 | {
571 | labels.append(line);
572 | line = textFile.readLine().trimmed();
573 | }
574 |
575 | textFile.close();
576 | }
577 | return true;
578 | }
579 | return false;
580 | }
581 |
582 | QString TensorFlow::getLabel(int index)
583 | {
584 | if(index>=0 && index=0 && index TensorFlow::getConfidence()
602 | {
603 | return rConfidence;
604 | }
605 |
606 | QList TensorFlow::getBoxes()
607 | {
608 | return rBox;
609 | }
610 |
611 | QList TensorFlow::getMasks()
612 | {
613 | return rMasks;
614 | }
615 |
616 | int TensorFlow::getInferenceTime()
617 | {
618 | return inferenceTime;
619 | }
620 |
621 | double TensorFlow::getResultConfidence(int index)
622 | {
623 | if (index>=0 && index
5 | Everyone is permitted to copy and distribute verbatim copies
6 | of this license document, but changing it is not allowed.
7 |
8 | Preamble
9 |
10 | The GNU General Public License is a free, copyleft license for
11 | software and other kinds of works.
12 |
13 | The licenses for most software and other practical works are designed
14 | to take away your freedom to share and change the works. By contrast,
15 | the GNU General Public License is intended to guarantee your freedom to
16 | share and change all versions of a program--to make sure it remains free
17 | software for all its users. We, the Free Software Foundation, use the
18 | GNU General Public License for most of our software; it applies also to
19 | any other work released this way by its authors. You can apply it to
20 | your programs, too.
21 |
22 | When we speak of free software, we are referring to freedom, not
23 | price. Our General Public Licenses are designed to make sure that you
24 | have the freedom to distribute copies of free software (and charge for
25 | them if you wish), that you receive source code or can get it if you
26 | want it, that you can change the software or use pieces of it in new
27 | free programs, and that you know you can do these things.
28 |
29 | To protect your rights, we need to prevent others from denying you
30 | these rights or asking you to surrender the rights. Therefore, you have
31 | certain responsibilities if you distribute copies of the software, or if
32 | you modify it: responsibilities to respect the freedom of others.
33 |
34 | For example, if you distribute copies of such a program, whether
35 | gratis or for a fee, you must pass on to the recipients the same
36 | freedoms that you received. You must make sure that they, too, receive
37 | or can get the source code. And you must show them these terms so they
38 | know their rights.
39 |
40 | Developers that use the GNU GPL protect your rights with two steps:
41 | (1) assert copyright on the software, and (2) offer you this License
42 | giving you legal permission to copy, distribute and/or modify it.
43 |
44 | For the developers' and authors' protection, the GPL clearly explains
45 | that there is no warranty for this free software. For both users' and
46 | authors' sake, the GPL requires that modified versions be marked as
47 | changed, so that their problems will not be attributed erroneously to
48 | authors of previous versions.
49 |
50 | Some devices are designed to deny users access to install or run
51 | modified versions of the software inside them, although the manufacturer
52 | can do so. This is fundamentally incompatible with the aim of
53 | protecting users' freedom to change the software. The systematic
54 | pattern of such abuse occurs in the area of products for individuals to
55 | use, which is precisely where it is most unacceptable. Therefore, we
56 | have designed this version of the GPL to prohibit the practice for those
57 | products. If such problems arise substantially in other domains, we
58 | stand ready to extend this provision to those domains in future versions
59 | of the GPL, as needed to protect the freedom of users.
60 |
61 | Finally, every program is threatened constantly by software patents.
62 | States should not allow patents to restrict development and use of
63 | software on general-purpose computers, but in those that do, we wish to
64 | avoid the special danger that patents applied to a free program could
65 | make it effectively proprietary. To prevent this, the GPL assures that
66 | patents cannot be used to render the program non-free.
67 |
68 | The precise terms and conditions for copying, distribution and
69 | modification follow.
70 |
71 | TERMS AND CONDITIONS
72 |
73 | 0. Definitions.
74 |
75 | "This License" refers to version 3 of the GNU General Public License.
76 |
77 | "Copyright" also means copyright-like laws that apply to other kinds of
78 | works, such as semiconductor masks.
79 |
80 | "The Program" refers to any copyrightable work licensed under this
81 | License. Each licensee is addressed as "you". "Licensees" and
82 | "recipients" may be individuals or organizations.
83 |
84 | To "modify" a work means to copy from or adapt all or part of the work
85 | in a fashion requiring copyright permission, other than the making of an
86 | exact copy. The resulting work is called a "modified version" of the
87 | earlier work or a work "based on" the earlier work.
88 |
89 | A "covered work" means either the unmodified Program or a work based
90 | on the Program.
91 |
92 | To "propagate" a work means to do anything with it that, without
93 | permission, would make you directly or secondarily liable for
94 | infringement under applicable copyright law, except executing it on a
95 | computer or modifying a private copy. Propagation includes copying,
96 | distribution (with or without modification), making available to the
97 | public, and in some countries other activities as well.
98 |
99 | To "convey" a work means any kind of propagation that enables other
100 | parties to make or receive copies. Mere interaction with a user through
101 | a computer network, with no transfer of a copy, is not conveying.
102 |
103 | An interactive user interface displays "Appropriate Legal Notices"
104 | to the extent that it includes a convenient and prominently visible
105 | feature that (1) displays an appropriate copyright notice, and (2)
106 | tells the user that there is no warranty for the work (except to the
107 | extent that warranties are provided), that licensees may convey the
108 | work under this License, and how to view a copy of this License. If
109 | the interface presents a list of user commands or options, such as a
110 | menu, a prominent item in the list meets this criterion.
111 |
112 | 1. Source Code.
113 |
114 | The "source code" for a work means the preferred form of the work
115 | for making modifications to it. "Object code" means any non-source
116 | form of a work.
117 |
118 | A "Standard Interface" means an interface that either is an official
119 | standard defined by a recognized standards body, or, in the case of
120 | interfaces specified for a particular programming language, one that
121 | is widely used among developers working in that language.
122 |
123 | The "System Libraries" of an executable work include anything, other
124 | than the work as a whole, that (a) is included in the normal form of
125 | packaging a Major Component, but which is not part of that Major
126 | Component, and (b) serves only to enable use of the work with that
127 | Major Component, or to implement a Standard Interface for which an
128 | implementation is available to the public in source code form. A
129 | "Major Component", in this context, means a major essential component
130 | (kernel, window system, and so on) of the specific operating system
131 | (if any) on which the executable work runs, or a compiler used to
132 | produce the work, or an object code interpreter used to run it.
133 |
134 | The "Corresponding Source" for a work in object code form means all
135 | the source code needed to generate, install, and (for an executable
136 | work) run the object code and to modify the work, including scripts to
137 | control those activities. However, it does not include the work's
138 | System Libraries, or general-purpose tools or generally available free
139 | programs which are used unmodified in performing those activities but
140 | which are not part of the work. For example, Corresponding Source
141 | includes interface definition files associated with source files for
142 | the work, and the source code for shared libraries and dynamically
143 | linked subprograms that the work is specifically designed to require,
144 | such as by intimate data communication or control flow between those
145 | subprograms and other parts of the work.
146 |
147 | The Corresponding Source need not include anything that users
148 | can regenerate automatically from other parts of the Corresponding
149 | Source.
150 |
151 | The Corresponding Source for a work in source code form is that
152 | same work.
153 |
154 | 2. Basic Permissions.
155 |
156 | All rights granted under this License are granted for the term of
157 | copyright on the Program, and are irrevocable provided the stated
158 | conditions are met. This License explicitly affirms your unlimited
159 | permission to run the unmodified Program. The output from running a
160 | covered work is covered by this License only if the output, given its
161 | content, constitutes a covered work. This License acknowledges your
162 | rights of fair use or other equivalent, as provided by copyright law.
163 |
164 | You may make, run and propagate covered works that you do not
165 | convey, without conditions so long as your license otherwise remains
166 | in force. You may convey covered works to others for the sole purpose
167 | of having them make modifications exclusively for you, or provide you
168 | with facilities for running those works, provided that you comply with
169 | the terms of this License in conveying all material for which you do
170 | not control copyright. Those thus making or running the covered works
171 | for you must do so exclusively on your behalf, under your direction
172 | and control, on terms that prohibit them from making any copies of
173 | your copyrighted material outside their relationship with you.
174 |
175 | Conveying under any other circumstances is permitted solely under
176 | the conditions stated below. Sublicensing is not allowed; section 10
177 | makes it unnecessary.
178 |
179 | 3. Protecting Users' Legal Rights From Anti-Circumvention Law.
180 |
181 | No covered work shall be deemed part of an effective technological
182 | measure under any applicable law fulfilling obligations under article
183 | 11 of the WIPO copyright treaty adopted on 20 December 1996, or
184 | similar laws prohibiting or restricting circumvention of such
185 | measures.
186 |
187 | When you convey a covered work, you waive any legal power to forbid
188 | circumvention of technological measures to the extent such circumvention
189 | is effected by exercising rights under this License with respect to
190 | the covered work, and you disclaim any intention to limit operation or
191 | modification of the work as a means of enforcing, against the work's
192 | users, your or third parties' legal rights to forbid circumvention of
193 | technological measures.
194 |
195 | 4. Conveying Verbatim Copies.
196 |
197 | You may convey verbatim copies of the Program's source code as you
198 | receive it, in any medium, provided that you conspicuously and
199 | appropriately publish on each copy an appropriate copyright notice;
200 | keep intact all notices stating that this License and any
201 | non-permissive terms added in accord with section 7 apply to the code;
202 | keep intact all notices of the absence of any warranty; and give all
203 | recipients a copy of this License along with the Program.
204 |
205 | You may charge any price or no price for each copy that you convey,
206 | and you may offer support or warranty protection for a fee.
207 |
208 | 5. Conveying Modified Source Versions.
209 |
210 | You may convey a work based on the Program, or the modifications to
211 | produce it from the Program, in the form of source code under the
212 | terms of section 4, provided that you also meet all of these conditions:
213 |
214 | a) The work must carry prominent notices stating that you modified
215 | it, and giving a relevant date.
216 |
217 | b) The work must carry prominent notices stating that it is
218 | released under this License and any conditions added under section
219 | 7. This requirement modifies the requirement in section 4 to
220 | "keep intact all notices".
221 |
222 | c) You must license the entire work, as a whole, under this
223 | License to anyone who comes into possession of a copy. This
224 | License will therefore apply, along with any applicable section 7
225 | additional terms, to the whole of the work, and all its parts,
226 | regardless of how they are packaged. This License gives no
227 | permission to license the work in any other way, but it does not
228 | invalidate such permission if you have separately received it.
229 |
230 | d) If the work has interactive user interfaces, each must display
231 | Appropriate Legal Notices; however, if the Program has interactive
232 | interfaces that do not display Appropriate Legal Notices, your
233 | work need not make them do so.
234 |
235 | A compilation of a covered work with other separate and independent
236 | works, which are not by their nature extensions of the covered work,
237 | and which are not combined with it such as to form a larger program,
238 | in or on a volume of a storage or distribution medium, is called an
239 | "aggregate" if the compilation and its resulting copyright are not
240 | used to limit the access or legal rights of the compilation's users
241 | beyond what the individual works permit. Inclusion of a covered work
242 | in an aggregate does not cause this License to apply to the other
243 | parts of the aggregate.
244 |
245 | 6. Conveying Non-Source Forms.
246 |
247 | You may convey a covered work in object code form under the terms
248 | of sections 4 and 5, provided that you also convey the
249 | machine-readable Corresponding Source under the terms of this License,
250 | in one of these ways:
251 |
252 | a) Convey the object code in, or embodied in, a physical product
253 | (including a physical distribution medium), accompanied by the
254 | Corresponding Source fixed on a durable physical medium
255 | customarily used for software interchange.
256 |
257 | b) Convey the object code in, or embodied in, a physical product
258 | (including a physical distribution medium), accompanied by a
259 | written offer, valid for at least three years and valid for as
260 | long as you offer spare parts or customer support for that product
261 | model, to give anyone who possesses the object code either (1) a
262 | copy of the Corresponding Source for all the software in the
263 | product that is covered by this License, on a durable physical
264 | medium customarily used for software interchange, for a price no
265 | more than your reasonable cost of physically performing this
266 | conveying of source, or (2) access to copy the
267 | Corresponding Source from a network server at no charge.
268 |
269 | c) Convey individual copies of the object code with a copy of the
270 | written offer to provide the Corresponding Source. This
271 | alternative is allowed only occasionally and noncommercially, and
272 | only if you received the object code with such an offer, in accord
273 | with subsection 6b.
274 |
275 | d) Convey the object code by offering access from a designated
276 | place (gratis or for a charge), and offer equivalent access to the
277 | Corresponding Source in the same way through the same place at no
278 | further charge. You need not require recipients to copy the
279 | Corresponding Source along with the object code. If the place to
280 | copy the object code is a network server, the Corresponding Source
281 | may be on a different server (operated by you or a third party)
282 | that supports equivalent copying facilities, provided you maintain
283 | clear directions next to the object code saying where to find the
284 | Corresponding Source. Regardless of what server hosts the
285 | Corresponding Source, you remain obligated to ensure that it is
286 | available for as long as needed to satisfy these requirements.
287 |
288 | e) Convey the object code using peer-to-peer transmission, provided
289 | you inform other peers where the object code and Corresponding
290 | Source of the work are being offered to the general public at no
291 | charge under subsection 6d.
292 |
293 | A separable portion of the object code, whose source code is excluded
294 | from the Corresponding Source as a System Library, need not be
295 | included in conveying the object code work.
296 |
297 | A "User Product" is either (1) a "consumer product", which means any
298 | tangible personal property which is normally used for personal, family,
299 | or household purposes, or (2) anything designed or sold for incorporation
300 | into a dwelling. In determining whether a product is a consumer product,
301 | doubtful cases shall be resolved in favor of coverage. For a particular
302 | product received by a particular user, "normally used" refers to a
303 | typical or common use of that class of product, regardless of the status
304 | of the particular user or of the way in which the particular user
305 | actually uses, or expects or is expected to use, the product. A product
306 | is a consumer product regardless of whether the product has substantial
307 | commercial, industrial or non-consumer uses, unless such uses represent
308 | the only significant mode of use of the product.
309 |
310 | "Installation Information" for a User Product means any methods,
311 | procedures, authorization keys, or other information required to install
312 | and execute modified versions of a covered work in that User Product from
313 | a modified version of its Corresponding Source. The information must
314 | suffice to ensure that the continued functioning of the modified object
315 | code is in no case prevented or interfered with solely because
316 | modification has been made.
317 |
318 | If you convey an object code work under this section in, or with, or
319 | specifically for use in, a User Product, and the conveying occurs as
320 | part of a transaction in which the right of possession and use of the
321 | User Product is transferred to the recipient in perpetuity or for a
322 | fixed term (regardless of how the transaction is characterized), the
323 | Corresponding Source conveyed under this section must be accompanied
324 | by the Installation Information. But this requirement does not apply
325 | if neither you nor any third party retains the ability to install
326 | modified object code on the User Product (for example, the work has
327 | been installed in ROM).
328 |
329 | The requirement to provide Installation Information does not include a
330 | requirement to continue to provide support service, warranty, or updates
331 | for a work that has been modified or installed by the recipient, or for
332 | the User Product in which it has been modified or installed. Access to a
333 | network may be denied when the modification itself materially and
334 | adversely affects the operation of the network or violates the rules and
335 | protocols for communication across the network.
336 |
337 | Corresponding Source conveyed, and Installation Information provided,
338 | in accord with this section must be in a format that is publicly
339 | documented (and with an implementation available to the public in
340 | source code form), and must require no special password or key for
341 | unpacking, reading or copying.
342 |
343 | 7. Additional Terms.
344 |
345 | "Additional permissions" are terms that supplement the terms of this
346 | License by making exceptions from one or more of its conditions.
347 | Additional permissions that are applicable to the entire Program shall
348 | be treated as though they were included in this License, to the extent
349 | that they are valid under applicable law. If additional permissions
350 | apply only to part of the Program, that part may be used separately
351 | under those permissions, but the entire Program remains governed by
352 | this License without regard to the additional permissions.
353 |
354 | When you convey a copy of a covered work, you may at your option
355 | remove any additional permissions from that copy, or from any part of
356 | it. (Additional permissions may be written to require their own
357 | removal in certain cases when you modify the work.) You may place
358 | additional permissions on material, added by you to a covered work,
359 | for which you have or can give appropriate copyright permission.
360 |
361 | Notwithstanding any other provision of this License, for material you
362 | add to a covered work, you may (if authorized by the copyright holders of
363 | that material) supplement the terms of this License with terms:
364 |
365 | a) Disclaiming warranty or limiting liability differently from the
366 | terms of sections 15 and 16 of this License; or
367 |
368 | b) Requiring preservation of specified reasonable legal notices or
369 | author attributions in that material or in the Appropriate Legal
370 | Notices displayed by works containing it; or
371 |
372 | c) Prohibiting misrepresentation of the origin of that material, or
373 | requiring that modified versions of such material be marked in
374 | reasonable ways as different from the original version; or
375 |
376 | d) Limiting the use for publicity purposes of names of licensors or
377 | authors of the material; or
378 |
379 | e) Declining to grant rights under trademark law for use of some
380 | trade names, trademarks, or service marks; or
381 |
382 | f) Requiring indemnification of licensors and authors of that
383 | material by anyone who conveys the material (or modified versions of
384 | it) with contractual assumptions of liability to the recipient, for
385 | any liability that these contractual assumptions directly impose on
386 | those licensors and authors.
387 |
388 | All other non-permissive additional terms are considered "further
389 | restrictions" within the meaning of section 10. If the Program as you
390 | received it, or any part of it, contains a notice stating that it is
391 | governed by this License along with a term that is a further
392 | restriction, you may remove that term. If a license document contains
393 | a further restriction but permits relicensing or conveying under this
394 | License, you may add to a covered work material governed by the terms
395 | of that license document, provided that the further restriction does
396 | not survive such relicensing or conveying.
397 |
398 | If you add terms to a covered work in accord with this section, you
399 | must place, in the relevant source files, a statement of the
400 | additional terms that apply to those files, or a notice indicating
401 | where to find the applicable terms.
402 |
403 | Additional terms, permissive or non-permissive, may be stated in the
404 | form of a separately written license, or stated as exceptions;
405 | the above requirements apply either way.
406 |
407 | 8. Termination.
408 |
409 | You may not propagate or modify a covered work except as expressly
410 | provided under this License. Any attempt otherwise to propagate or
411 | modify it is void, and will automatically terminate your rights under
412 | this License (including any patent licenses granted under the third
413 | paragraph of section 11).
414 |
415 | However, if you cease all violation of this License, then your
416 | license from a particular copyright holder is reinstated (a)
417 | provisionally, unless and until the copyright holder explicitly and
418 | finally terminates your license, and (b) permanently, if the copyright
419 | holder fails to notify you of the violation by some reasonable means
420 | prior to 60 days after the cessation.
421 |
422 | Moreover, your license from a particular copyright holder is
423 | reinstated permanently if the copyright holder notifies you of the
424 | violation by some reasonable means, this is the first time you have
425 | received notice of violation of this License (for any work) from that
426 | copyright holder, and you cure the violation prior to 30 days after
427 | your receipt of the notice.
428 |
429 | Termination of your rights under this section does not terminate the
430 | licenses of parties who have received copies or rights from you under
431 | this License. If your rights have been terminated and not permanently
432 | reinstated, you do not qualify to receive new licenses for the same
433 | material under section 10.
434 |
435 | 9. Acceptance Not Required for Having Copies.
436 |
437 | You are not required to accept this License in order to receive or
438 | run a copy of the Program. Ancillary propagation of a covered work
439 | occurring solely as a consequence of using peer-to-peer transmission
440 | to receive a copy likewise does not require acceptance. However,
441 | nothing other than this License grants you permission to propagate or
442 | modify any covered work. These actions infringe copyright if you do
443 | not accept this License. Therefore, by modifying or propagating a
444 | covered work, you indicate your acceptance of this License to do so.
445 |
446 | 10. Automatic Licensing of Downstream Recipients.
447 |
448 | Each time you convey a covered work, the recipient automatically
449 | receives a license from the original licensors, to run, modify and
450 | propagate that work, subject to this License. You are not responsible
451 | for enforcing compliance by third parties with this License.
452 |
453 | An "entity transaction" is a transaction transferring control of an
454 | organization, or substantially all assets of one, or subdividing an
455 | organization, or merging organizations. If propagation of a covered
456 | work results from an entity transaction, each party to that
457 | transaction who receives a copy of the work also receives whatever
458 | licenses to the work the party's predecessor in interest had or could
459 | give under the previous paragraph, plus a right to possession of the
460 | Corresponding Source of the work from the predecessor in interest, if
461 | the predecessor has it or can get it with reasonable efforts.
462 |
463 | You may not impose any further restrictions on the exercise of the
464 | rights granted or affirmed under this License. For example, you may
465 | not impose a license fee, royalty, or other charge for exercise of
466 | rights granted under this License, and you may not initiate litigation
467 | (including a cross-claim or counterclaim in a lawsuit) alleging that
468 | any patent claim is infringed by making, using, selling, offering for
469 | sale, or importing the Program or any portion of it.
470 |
471 | 11. Patents.
472 |
473 | A "contributor" is a copyright holder who authorizes use under this
474 | License of the Program or a work on which the Program is based. The
475 | work thus licensed is called the contributor's "contributor version".
476 |
477 | A contributor's "essential patent claims" are all patent claims
478 | owned or controlled by the contributor, whether already acquired or
479 | hereafter acquired, that would be infringed by some manner, permitted
480 | by this License, of making, using, or selling its contributor version,
481 | but do not include claims that would be infringed only as a
482 | consequence of further modification of the contributor version. For
483 | purposes of this definition, "control" includes the right to grant
484 | patent sublicenses in a manner consistent with the requirements of
485 | this License.
486 |
487 | Each contributor grants you a non-exclusive, worldwide, royalty-free
488 | patent license under the contributor's essential patent claims, to
489 | make, use, sell, offer for sale, import and otherwise run, modify and
490 | propagate the contents of its contributor version.
491 |
492 | In the following three paragraphs, a "patent license" is any express
493 | agreement or commitment, however denominated, not to enforce a patent
494 | (such as an express permission to practice a patent or covenant not to
495 | sue for patent infringement). To "grant" such a patent license to a
496 | party means to make such an agreement or commitment not to enforce a
497 | patent against the party.
498 |
499 | If you convey a covered work, knowingly relying on a patent license,
500 | and the Corresponding Source of the work is not available for anyone
501 | to copy, free of charge and under the terms of this License, through a
502 | publicly available network server or other readily accessible means,
503 | then you must either (1) cause the Corresponding Source to be so
504 | available, or (2) arrange to deprive yourself of the benefit of the
505 | patent license for this particular work, or (3) arrange, in a manner
506 | consistent with the requirements of this License, to extend the patent
507 | license to downstream recipients. "Knowingly relying" means you have
508 | actual knowledge that, but for the patent license, your conveying the
509 | covered work in a country, or your recipient's use of the covered work
510 | in a country, would infringe one or more identifiable patents in that
511 | country that you have reason to believe are valid.
512 |
513 | If, pursuant to or in connection with a single transaction or
514 | arrangement, you convey, or propagate by procuring conveyance of, a
515 | covered work, and grant a patent license to some of the parties
516 | receiving the covered work authorizing them to use, propagate, modify
517 | or convey a specific copy of the covered work, then the patent license
518 | you grant is automatically extended to all recipients of the covered
519 | work and works based on it.
520 |
521 | A patent license is "discriminatory" if it does not include within
522 | the scope of its coverage, prohibits the exercise of, or is
523 | conditioned on the non-exercise of one or more of the rights that are
524 | specifically granted under this License. You may not convey a covered
525 | work if you are a party to an arrangement with a third party that is
526 | in the business of distributing software, under which you make payment
527 | to the third party based on the extent of your activity of conveying
528 | the work, and under which the third party grants, to any of the
529 | parties who would receive the covered work from you, a discriminatory
530 | patent license (a) in connection with copies of the covered work
531 | conveyed by you (or copies made from those copies), or (b) primarily
532 | for and in connection with specific products or compilations that
533 | contain the covered work, unless you entered into that arrangement,
534 | or that patent license was granted, prior to 28 March 2007.
535 |
536 | Nothing in this License shall be construed as excluding or limiting
537 | any implied license or other defenses to infringement that may
538 | otherwise be available to you under applicable patent law.
539 |
540 | 12. No Surrender of Others' Freedom.
541 |
542 | If conditions are imposed on you (whether by court order, agreement or
543 | otherwise) that contradict the conditions of this License, they do not
544 | excuse you from the conditions of this License. If you cannot convey a
545 | covered work so as to satisfy simultaneously your obligations under this
546 | License and any other pertinent obligations, then as a consequence you may
547 | not convey it at all. For example, if you agree to terms that obligate you
548 | to collect a royalty for further conveying from those to whom you convey
549 | the Program, the only way you could satisfy both those terms and this
550 | License would be to refrain entirely from conveying the Program.
551 |
552 | 13. Use with the GNU Affero General Public License.
553 |
554 | Notwithstanding any other provision of this License, you have
555 | permission to link or combine any covered work with a work licensed
556 | under version 3 of the GNU Affero General Public License into a single
557 | combined work, and to convey the resulting work. The terms of this
558 | License will continue to apply to the part which is the covered work,
559 | but the special requirements of the GNU Affero General Public License,
560 | section 13, concerning interaction through a network will apply to the
561 | combination as such.
562 |
563 | 14. Revised Versions of this License.
564 |
565 | The Free Software Foundation may publish revised and/or new versions of
566 | the GNU General Public License from time to time. Such new versions will
567 | be similar in spirit to the present version, but may differ in detail to
568 | address new problems or concerns.
569 |
570 | Each version is given a distinguishing version number. If the
571 | Program specifies that a certain numbered version of the GNU General
572 | Public License "or any later version" applies to it, you have the
573 | option of following the terms and conditions either of that numbered
574 | version or of any later version published by the Free Software
575 | Foundation. If the Program does not specify a version number of the
576 | GNU General Public License, you may choose any version ever published
577 | by the Free Software Foundation.
578 |
579 | If the Program specifies that a proxy can decide which future
580 | versions of the GNU General Public License can be used, that proxy's
581 | public statement of acceptance of a version permanently authorizes you
582 | to choose that version for the Program.
583 |
584 | Later license versions may give you additional or different
585 | permissions. However, no additional obligations are imposed on any
586 | author or copyright holder as a result of your choosing to follow a
587 | later version.
588 |
589 | 15. Disclaimer of Warranty.
590 |
591 | THERE IS NO WARRANTY FOR THE PROGRAM, TO THE EXTENT PERMITTED BY
592 | APPLICABLE LAW. EXCEPT WHEN OTHERWISE STATED IN WRITING THE COPYRIGHT
593 | HOLDERS AND/OR OTHER PARTIES PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY
594 | OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO,
595 | THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
596 | PURPOSE. THE ENTIRE RISK AS TO THE QUALITY AND PERFORMANCE OF THE PROGRAM
597 | IS WITH YOU. SHOULD THE PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF
598 | ALL NECESSARY SERVICING, REPAIR OR CORRECTION.
599 |
600 | 16. Limitation of Liability.
601 |
602 | IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING
603 | WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MODIFIES AND/OR CONVEYS
604 | THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY
605 | GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING OUT OF THE
606 | USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED TO LOSS OF
607 | DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY YOU OR THIRD
608 | PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER PROGRAMS),
609 | EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE POSSIBILITY OF
610 | SUCH DAMAGES.
611 |
612 | 17. Interpretation of Sections 15 and 16.
613 |
614 | If the disclaimer of warranty and limitation of liability provided
615 | above cannot be given local legal effect according to their terms,
616 | reviewing courts shall apply local law that most closely approximates
617 | an absolute waiver of all civil liability in connection with the
618 | Program, unless a warranty or assumption of liability accompanies a
619 | copy of the Program in return for a fee.
620 |
621 | END OF TERMS AND CONDITIONS
622 |
623 | How to Apply These Terms to Your New Programs
624 |
625 | If you develop a new program, and you want it to be of the greatest
626 | possible use to the public, the best way to achieve this is to make it
627 | free software which everyone can redistribute and change under these terms.
628 |
629 | To do so, attach the following notices to the program. It is safest
630 | to attach them to the start of each source file to most effectively
631 | state the exclusion of warranty; and each file should have at least
632 | the "copyright" line and a pointer to where the full notice is found.
633 |
634 |
635 | Copyright (C)
636 |
637 | This program is free software: you can redistribute it and/or modify
638 | it under the terms of the GNU General Public License as published by
639 | the Free Software Foundation, either version 3 of the License, or
640 | (at your option) any later version.
641 |
642 | This program is distributed in the hope that it will be useful,
643 | but WITHOUT ANY WARRANTY; without even the implied warranty of
644 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
645 | GNU General Public License for more details.
646 |
647 | You should have received a copy of the GNU General Public License
648 | along with this program. If not, see .
649 |
650 | Also add information on how to contact you by electronic and paper mail.
651 |
652 | If the program does terminal interaction, make it output a short
653 | notice like this when it starts in an interactive mode:
654 |
655 | Copyright (C)
656 | This program comes with ABSOLUTELY NO WARRANTY; for details type `show w'.
657 | This is free software, and you are welcome to redistribute it
658 | under certain conditions; type `show c' for details.
659 |
660 | The hypothetical commands `show w' and `show c' should show the appropriate
661 | parts of the General Public License. Of course, your program's commands
662 | might be different; for a GUI interface, you would use an "about box".
663 |
664 | You should also get your employer (if you work as a programmer) or school,
665 | if any, to sign a "copyright disclaimer" for the program, if necessary.
666 | For more information on this, and how to apply and follow the GNU GPL, see
667 | .
668 |
669 | The GNU General Public License does not permit incorporating your program
670 | into proprietary programs. If your program is a subroutine library, you
671 | may consider it more useful to permit linking proprietary applications with
672 | the library. If this is what you want to do, use the GNU Lesser General
673 | Public License instead of this License. But first, please read
674 | .
675 |
--------------------------------------------------------------------------------
/Pi_Qt_TFLite_DeepLab.pro.user.4.9-pre1:
--------------------------------------------------------------------------------
1 |
2 |
3 |
4 |
5 |
6 | EnvironmentId
7 | {fb7d6e16-c15b-4f01-98f9-9e14e4583b5b}
8 |
9 |
10 | ProjectExplorer.Project.ActiveTarget
11 | 0
12 |
13 |
14 | ProjectExplorer.Project.EditorSettings
15 |
16 | true
17 | false
18 | true
19 |
20 | Cpp
21 |
22 | CppGlobal
23 |
24 |
25 |
26 | QmlJS
27 |
28 | QmlJSGlobal
29 |
30 |
31 | 2
32 | UTF-8
33 | false
34 | 4
35 | false
36 | 80
37 | true
38 | true
39 | 1
40 | true
41 | false
42 | 0
43 | true
44 | true
45 | 0
46 | 8
47 | true
48 | 1
49 | true
50 | true
51 | true
52 | false
53 |
54 |
55 |
56 | ProjectExplorer.Project.PluginSettings
57 |
58 |
59 |
60 | ProjectExplorer.Project.Target.0
61 |
62 | Desktop
63 | Desktop
64 | {e92845a8-1d9a-40ad-88a4-2073a741b729}
65 | 0
66 | 0
67 | 0
68 |
69 | /home/javi/Qt/build-Pi_Qt_TFLite_DeepLab-Desktop-Debug
70 |
71 |
72 | true
73 | qmake
74 |
75 | QtProjectManager.QMakeBuildStep
76 | true
77 |
78 | false
79 | false
80 | false
81 |
82 |
83 | true
84 | Make
85 |
86 | Qt4ProjectManager.MakeStep
87 |
88 | false
89 |
90 |
91 | false
92 |
93 | 2
94 | Build
95 |
96 | ProjectExplorer.BuildSteps.Build
97 |
98 |
99 |
100 | true
101 | Make
102 |
103 | Qt4ProjectManager.MakeStep
104 |
105 | true
106 | clean
107 |
108 | false
109 |
110 | 1
111 | Clean
112 |
113 | ProjectExplorer.BuildSteps.Clean
114 |
115 | 2
116 | false
117 |
118 | Debug
119 | Debug
120 | Qt4ProjectManager.Qt4BuildConfiguration
121 | 2
122 | true
123 |
124 |
125 | /home/javi/Qt/build-Pi_Qt_TFLite_DeepLab-Desktop-Release
126 |
127 |
128 | true
129 | qmake
130 |
131 | QtProjectManager.QMakeBuildStep
132 | false
133 |
134 | false
135 | false
136 | true
137 |
138 |
139 | true
140 | Make
141 |
142 | Qt4ProjectManager.MakeStep
143 |
144 | false
145 |
146 |
147 | false
148 |
149 | 2
150 | Build
151 |
152 | ProjectExplorer.BuildSteps.Build
153 |
154 |
155 |
156 | true
157 | Make
158 |
159 | Qt4ProjectManager.MakeStep
160 |
161 | true
162 | clean
163 |
164 | false
165 |
166 | 1
167 | Clean
168 |
169 | ProjectExplorer.BuildSteps.Clean
170 |
171 | 2
172 | false
173 |
174 | Release
175 | Release
176 | Qt4ProjectManager.Qt4BuildConfiguration
177 | 0
178 | true
179 |
180 |
181 | /home/javi/Qt/build-Pi_Qt_TFLite_DeepLab-Desktop-Profile
182 |
183 |
184 | true
185 | qmake
186 |
187 | QtProjectManager.QMakeBuildStep
188 | true
189 |
190 | false
191 | true
192 | true
193 |
194 |
195 | true
196 | Make
197 |
198 | Qt4ProjectManager.MakeStep
199 |
200 | false
201 |
202 |
203 | false
204 |
205 | 2
206 | Build
207 |
208 | ProjectExplorer.BuildSteps.Build
209 |
210 |
211 |
212 | true
213 | Make
214 |
215 | Qt4ProjectManager.MakeStep
216 |
217 | true
218 | clean
219 |
220 | false
221 |
222 | 1
223 | Clean
224 |
225 | ProjectExplorer.BuildSteps.Clean
226 |
227 | 2
228 | false
229 |
230 | Profile
231 | Profile
232 | Qt4ProjectManager.Qt4BuildConfiguration
233 | 0
234 | true
235 |
236 | 3
237 |
238 |
239 | 0
240 | Deploy
241 |
242 | ProjectExplorer.BuildSteps.Deploy
243 |
244 | 1
245 | Deploy Configuration
246 |
247 | ProjectExplorer.DefaultDeployConfiguration
248 |
249 | 1
250 |
251 |
252 | false
253 | false
254 | 1000
255 |
256 | true
257 |
258 | false
259 | false
260 | false
261 | false
262 | true
263 | 0.01
264 | 10
265 | true
266 | 1
267 | 25
268 |
269 | 1
270 | true
271 | false
272 | true
273 | valgrind
274 |
275 | 0
276 | 1
277 | 2
278 | 3
279 | 4
280 | 5
281 | 6
282 | 7
283 | 8
284 | 9
285 | 10
286 | 11
287 | 12
288 | 13
289 | 14
290 |
291 | 2
292 |
293 | Pi_Qt_TFLite_DeepLab
294 |
295 | Qt4ProjectManager.Qt4RunConfiguration:/home/javi/Qt/Pi_Qt_TFLite_DeepLab/Pi_Qt_TFLite_DeepLab.pro
296 | Pi_Qt_TFLite_DeepLab.pro
297 |
298 | 3768
299 | false
300 | true
301 | true
302 | false
303 | false
304 | true
305 |
306 | /home/javi/Qt/build-Pi_Qt_TFLite_DeepLab-Desktop-Debug
307 |
308 | 1
309 |
310 |
311 |
312 | ProjectExplorer.Project.Target.1
313 |
314 | Raspberry Pi
315 | Raspberry Pi
316 | {e8fded20-8fae-4073-a335-926c5fc16d63}
317 | 0
318 | 0
319 | 0
320 |
321 | /home/javi/Qt/build-Pi_Qt_TFLite_DeepLab-Raspberry_Pi-Debug
322 |
323 |
324 | true
325 | qmake
326 |
327 | QtProjectManager.QMakeBuildStep
328 | true
329 |
330 | false
331 | false
332 | false
333 |
334 |
335 | true
336 | Make
337 |
338 | Qt4ProjectManager.MakeStep
339 |
340 | false
341 |
342 |
343 | false
344 |
345 | 2
346 | Build
347 |
348 | ProjectExplorer.BuildSteps.Build
349 |
350 |
351 |
352 | true
353 | Make
354 |
355 | Qt4ProjectManager.MakeStep
356 |
357 | true
358 | clean
359 |
360 | false
361 |
362 | 1
363 | Clean
364 |
365 | ProjectExplorer.BuildSteps.Clean
366 |
367 | 2
368 | false
369 |
370 | Debug
371 | Debug
372 | Qt4ProjectManager.Qt4BuildConfiguration
373 | 2
374 | true
375 |
376 |
377 | /home/javi/Qt/build-Pi_Qt_TFLite_DeepLab-Raspberry_Pi-Release
378 |
379 |
380 | true
381 | qmake
382 |
383 | QtProjectManager.QMakeBuildStep
384 | false
385 |
386 | false
387 | false
388 | true
389 |
390 |
391 | true
392 | Make
393 |
394 | Qt4ProjectManager.MakeStep
395 |
396 | false
397 |
398 |
399 | false
400 |
401 | 2
402 | Build
403 |
404 | ProjectExplorer.BuildSteps.Build
405 |
406 |
407 |
408 | true
409 | Make
410 |
411 | Qt4ProjectManager.MakeStep
412 |
413 | true
414 | clean
415 |
416 | false
417 |
418 | 1
419 | Clean
420 |
421 | ProjectExplorer.BuildSteps.Clean
422 |
423 | 2
424 | false
425 |
426 | Release
427 | Release
428 | Qt4ProjectManager.Qt4BuildConfiguration
429 | 0
430 | true
431 |
432 |
433 | /home/javi/Qt/build-Pi_Qt_TFLite_DeepLab-Raspberry_Pi-Profile
434 |
435 |
436 | true
437 | qmake
438 |
439 | QtProjectManager.QMakeBuildStep
440 | true
441 |
442 | false
443 | true
444 | true
445 |
446 |
447 | true
448 | Make
449 |
450 | Qt4ProjectManager.MakeStep
451 |
452 | false
453 |
454 |
455 | false
456 |
457 | 2
458 | Build
459 |
460 | ProjectExplorer.BuildSteps.Build
461 |
462 |
463 |
464 | true
465 | Make
466 |
467 | Qt4ProjectManager.MakeStep
468 |
469 | true
470 | clean
471 |
472 | false
473 |
474 | 1
475 | Clean
476 |
477 | ProjectExplorer.BuildSteps.Clean
478 |
479 | 2
480 | false
481 |
482 | Profile
483 | Profile
484 | Qt4ProjectManager.Qt4BuildConfiguration
485 | 0
486 | true
487 |
488 | 3
489 |
490 |
491 |
492 | true
493 | Check for free disk space
494 |
495 | RemoteLinux.CheckForFreeDiskSpaceStep
496 |
497 |
498 |
499 |
500 |
501 | /
502 | 5242880
503 |
504 |
505 | true
506 | Upload files via SFTP
507 |
508 | RemoteLinux.DirectUploadStep
509 |
510 | /home/javi/Qt/build-Pi_Qt_TFLite_DeepLab-Raspberry_Pi-Debug/Pi_Qt_TFLite_DeepLab
511 | /home/javi/Qt/Pi_Qt_TFLite_DeepLab/assets/deeplabv3_257_mv_gpu.tflite
512 |
513 |
514 | 192.168.1.45
515 | 192.168.1.45
516 |
517 |
518 | /home/pi/qt_apps/Pi_Qt_TFLite_DeepLab/bin
519 | /home/pi/qt_apps/Pi_Qt_TFLite_DeepLab/bin/assets
520 |
521 |
522 | /home/javi/raspi/sysroot
523 | /home/javi/raspi/sysroot
524 |
525 |
526 | 2019-05-01T14:08:11.748
527 | 2019-05-01T13:47:58.146
528 |
529 | false
530 | true
531 |
532 |
533 | true
534 | Kill current application instance
535 |
536 | RemoteLinux.KillAppStep
537 |
538 |
539 |
540 |
541 |
542 |
543 | 3
544 | Deploy
545 |
546 | ProjectExplorer.BuildSteps.Deploy
547 |
548 | 1
549 | Deploy to Remote Linux Host
550 |
551 | DeployToGenericLinux
552 |
553 | 1
554 |
555 |
556 | false
557 | false
558 | 1000
559 |
560 | true
561 |
562 | false
563 | false
564 | false
565 | false
566 | true
567 | 0.01
568 | 10
569 | true
570 | 1
571 | 25
572 |
573 | 1
574 | true
575 | false
576 | true
577 | valgrind
578 |
579 | 0
580 | 1
581 | 2
582 | 3
583 | 4
584 | 5
585 | 6
586 | 7
587 | 8
588 | 9
589 | 10
590 | 11
591 | 12
592 | 13
593 | 14
594 |
595 | 1
596 |
597 | Pi_Qt_TFLite_DeepLab (on Raspberry Pi)
598 |
599 | RemoteLinuxRunConfiguration:/home/javi/Qt/Pi_Qt_TFLite_DeepLab/Pi_Qt_TFLite_DeepLab.pro
600 | 1
601 |
602 | false
603 |
604 | 3768
605 | false
606 | true
607 | false
608 | false
609 | true
610 | /home/pi/qt_apps/Pi_Qt_TFLite_DeepLab/bin/
611 |
612 |
613 | 1
614 |
615 |
616 |
617 | ProjectExplorer.Project.TargetCount
618 | 2
619 |
620 |
621 | ProjectExplorer.Project.Updater.FileVersion
622 | 20
623 |
624 |
625 | Version
626 | 20
627 |
628 |
629 |
--------------------------------------------------------------------------------