Sie sind auf Seite 1von 9

Thread Management

SHEEZA SHAH
KINZA FATIMA
LAIBA SHAHID

Department of Computer Science


Bahria University, Karachi Campus

December 2019

1
INTRODUCTION
Threads represent a software approach to improving performance of operating system by
reducing the overhead thread is equivalent to a classical process. Each thread belongs to exactly
one process and no thread can exist outside a process. Each thread represents a separate flow of
control. In computer science, a thread of execution is the smallest sequence of programmed
instructions that can be managed independently by a scheduler, which is typically a part of
the operating system.[1] The implementation of threads and processes differs between operating
systems, but in most cases a thread is a component of a process. Multiple threads can exist within
one process, executing concurrently and sharing resources such as memory, while different
processes do not share these resources.

OBJECTIVES

 Describe thread management.


 Use threads in applications.
 Use threads with Python library functions.
 Build and execute threaded applications.

OUTCOME

 Multithreaded programs can run faster on computer systems with multiple CPUs, because
theses threads can be executed truly concurrent.
 A program can remain responsive to input. This is true both on single and on multiple
CPU
 Threads of a process can share the memory of global variables. If a global variable is
changed in one thread, this change is valid for all threads. A thread can have local
variables.

2
Final Deliverable of the Project:
There are two modules which support the usage of threads in Python:
 thread
and
 threading

BACKGROUND/LITERATURE REVIEW

Few areas of programming are more filled with myths than thread management is. Threads can become a
nightmare if you take an undisciplined approach, because once you start using threads you’re immediately
at the mercy of the operating system scheduler. Your thread code may suffer starvation, or you may discover
inadvertently that you’ve written code that isn’t thread-safe. Or you may find that your code gets itself into
a deadlock.
All of these problems can be avoided if you apply a solid design approach, which gives you all the
considerable power of multithreaded code. This article has shown how a long-running, complex
software upload procedure can be made responsive and asynchronous—all with just a few lines of
code!
There’s absolutely no reason why you shouldn’t incorporate multithreaded code into your own
efforts. It’s a powerful mechanism, and the platforms are making it easier than ever to achieve.

3
PROJECT METHODOLOGY

Video stream image segregator:

The video stream is constantly recorded through the on-board camera of the drone, and then
segmented into sequences of still images

Hand gesture recognition:

Each image is then analyzed through the hand gesture recognition process, which
includes three main steps:

 feature extraction
 hand region identification
 finally gesture classification

Command mapper:

A command mapper transforms the detected gesture into a command, such as takeoff,
land, or back off.

4
Action planner:

An action planner takes the command as its input and compute the corresponding
course of primitive actions to satisfy the command. While the planner is operating, it
also considers the surrounding environment to avoid collision and ensure the safety for
both the drone and perceived obstacles.

The set of all five gestures includes fist, palm, go symbol, V-shape, and little finger.

In order to implement the complete framework, there are a number of key challenges
that we need to address, namely gesture recognition, visual variability of scene, and
safety assurance of maneuver.

Visual Variability of Scene:

The proposed study is designed for a user to control a drone in daily life, not a special
laboratory environment. For this very reason, we want to empirically measure the
effects of scene variability while classification framework is kept unchanged

Safety:

After a gesture is recognized and converted to a command, such as move to the left,
the action planner on the drone kicks in to compute the most appropriate action. In this
process, it is imperative for the drone to carry out the action while ensuring safety to
itself, surrounding objects and environment

5
PROJECT SCHEDULE

KEY MILESTONES
Key Milestones of the Project with dates
S. No Elapsed time since start of the project Milestone Deliverable
6 hours/day (7 days) Project Title Research
1
Selection regarding topic.
2
3
4

6
GANTT CHART

RESOURSE ESTIMATION

Equipment required for making prototype/working model:


EQUIPMENT QUANTITY PRICES

Drone 1

Sony Spresense boards 1 $129.00


(main & extension)
Gesture Sensor 1 $14.95

Raspberry Pi3 Model B 1 $37.5


SparkFun Pushbutton switch 4 $4.48
12mm

Arduino UNO 1
Leap motion controller 1

SD card 1

Receiver and transmitter 1

Microcontroller 1

7
PCF8519P 1

 Snappy Ubuntu Core


 Parrot SDK
Softwares  Jupyter Notebook
 NetBeans IDE 7.1
 AIR FLASH
 Serialization

Languages  Python
 Javascript
 Node.js
 Leap.js

8
1. D
REFERENCES

[1] Priya Matnani, “GLOVE BASED AND ACCELEROMETER BASED GESTURE


CONTROL: A LITERATURE REVIEW”, International Journal of Technical Research and
Applications, www.ijtra.com Volume 3, Issue 6 (November-December, 2015), PP. 216-221

[2] “HAND GESTURE CONTROLLED AERIAL SURVEILLANCE DRONE”, International


Journal of Pure and Applied Mathematics Volume 119 No. 15 2018, 897-901 ISSN: 1314-3395
(on-line version), http://www.acadpubl.eu/hub/

[3] Ms. Tingare B.A. Mrs. Kolhe V. L.,” Controlling the Drone with Hand Gestures by using LEAP
Motion Controller”, International Journal of Pure and Applied Mathematics Volume 118 No. 24
2018 ISSN: 1314-3395 (on-line version) url: http://www.acadpubl.eu/hub/

[4] Kathiravan Natarajan ,Truong-Huy D. Nguyen and Mutlu Mete

“HandGestureControlledDrones:AnOpenSourceLibrary”

[5] Parrot, “Parrot ar.drone 2.0 power edition,” Nov 2017. [Online]. Available:
https://www.parrot.com/us/drones/parrot-ardrone20-power-edition

[6] Guoqing-Meng and M. Wang, “Hand gesture recognition based on fingertip detection,” in 2013
Fourth Global Congress on Intelligent Systems, Dec 2013, pp. 107–111.

[7] Moniruzzaman Bhuiyan and Rich Picking . ‘Gesturecontrolled user interfaces”. Centre for
Applied Internet Research (CAIR), Glyndŵr University, Wrexham, UK

[8]SMCorsham, ‘MyWingman: Gesture Controlled Drone”,Louborough University and NI

Das könnte Ihnen auch gefallen