• No results found

PERSONAL INTELLIGENT ASSISTANT FOR SPEECH RECOGNITION

N/A
N/A
Protected

Academic year: 2020

Share "PERSONAL INTELLIGENT ASSISTANT FOR SPEECH RECOGNITION"

Copied!
6
0
0

Loading.... (view fulltext now)

Full text

(1)

Personal Intelligent Assistant For Speech Recognition

Doniya Antony1, Jancy M P2 and Satheesh Kumar R3

1,2Student, Department of Computer Science and Engineering, Sahrdaya College of Engineering and Technology, Kerala, India.

3Associate Professor, Department of Computer Science and Engineering, Sahrdaya College of Engineering and Technology, Kerala, India.

Article Received: 27 November 2017 Article Accepted: 24 January 2018 Article Published: 2 March 2018

1. INTRODUCTION

Dialect is man's most imperative methods for correspondence and discourse its essential medium. Here we manage

this communication between the man and machine through union and acknowledgment applications. Speech

Recognition, includes catching and digitizing the sound waves, changing over them to fundamental dialect units,

building words from this units, and logically breaking down the words to guarantee adjust spelling for words that

sound alike. Speech Recognition is the capacity of a PC to perceive general, normally streaming articulations from

a wide assortment of clients. Speech Recognition enables you to furnish contribution to an application with your

voice.

Speech Recognition innovation is one from the quickly developing building advances. It has various applications in

various regions and gives potential advantages. Almost 20% individuals of the world are experiencing different

inabilities; a large number of them are visually impaired or unfit to utilize their hands adequately. The speech

recognition frameworks in those specific cases give a noteworthy help to them, so they can impart data to

individuals by working PC through voice input. It is an individual right hand for those peoples. The sound waves

are caught utilizing a recorder and it is associated with Rasperry Pi, it resembles a computer. The recorded voice is

tested and given as contribution to the framework to change over into content utilizing discourse to content

converter.

At that point the content is sought utilizing a Wi Fi and discovers its answer from Amazon Web server. The

recovered answer is on the other hand changed over back to speech utilizing content to speech converter. The

framework can be prepared by human. It utilizes certain classifiers to perceive the speech with most extreme

A B S T R A C T

Now we are in digital era. The technology is increasing rapidly and the knowledge turnover is high. It is out of the control of humans and becomes difficult to handle. So the knowledge spreading and availability should be equal for everyone. If we having any doubts regarding any query we can just google out. But this is complex task for those like visual impairment patients, slow typist etc. To fill this gap, we propose a technique to aid visual impairment persons, slow typist to enable to search over internet. In our system we can just ask our doubts, regarding any query, the system will talk to you by replying the answer. This is a smart machine which is working using a smart low cost SBC called Raspberry Pi. It uses the new methods of speech recognition to make the system to be more efficient. There are many existing systems like Apple’s siri, S voice etc. But these systems are privately owned. In order to make it accessible to common people, we provide a portable system at low cost which the public can easily access and use according to their needs. It is connected using the Wi Fi or Ethernet to retrieve the answers for the queries.

(2)

2.RELATEDWORKS

There are some works related to our work. They are mainly

Apple’s Siri, S voice, Google Now.

Apple’s Siri is the Speech Interpretation and Recognition Interface used in watchOS, macOS and tv OS operating

systems. It work as an intelligent personal assistant using some softwares. Siri can perform various user commands

in the system. Moreover it is able to engage with iOS integrated applications. The main advantage is usage of

natural language. The application captures the speech given by the speaker and then it will perform the narrated

function. The languages available in Siri is Arabic, English, French, Spanish etc.

S voice is a virtual mobile intelligent personal assistant used only in Samsung Smart phones. It is a knowledge

navigator. It can run large number of tasks through voice commands. The advantage is that it saves time and effort

to perform those functions from typing. This voice assistant is preloaded in the system. The application uses

mainly natural language interface to answer questions and make recommendation and perform actions by

delegating request to set of web services.

Google Now is the feature of Google Search in the google app for Android and iOS. It is a google product that

allows the users to search over internet using voice. It is a device search for data upon entering information on

what to search in the internet into the device by speaking. Now it is updated and the Google can talk back to us.

Earlier the queries were asked and the answers were getting through text through displaying on the screen.

Iris is a personal assistant application for Android. The application uses natural language processing to answer

questions based on user voice request. It is developed by Dextra Software Solutions. Iris can talk on topics ranging

from Philosophy, Culture, History, science to general conversation.

3. SYSTEM DESIGN AND IMPLEMENTATION

The basic hardware components used in our project are.

1. Raspberry Pi

2. Sound card

3. Microphone

4. Headset

5. Ethernet Cable

6. Memory Card

7. LED and switches.

(3)

Fig 1. System Overview

Fig.1 Capturing voice query

Initially audio is input the system. The audio will be sampled and will convert to text by STT conversion methods.

This text will search over internet and will find the answer to the query. This answer will again translated to speech

and give to the user.

The system consists of four functional modules:

1. Shell Script Voice Capturing

2. Google API for STT Conversion

3. Wolfram Alpha Engine API Interface

4. TTS Module Interface

Shell Script Voice Capturing

(4)

Basic shell contents are basically a grouping of basic orders

Google API for STT Conversion

For the caught speech, it is then to be expected to change over as a content configuration. For this reason Google

STT API is utilized. A decent favorable position of Google STT is that, there will no issues with respect to the

accents. The API perceives more than 110 dialects and variations, to help worldwide client base. We can likewise

channel unseemly substance in content outcomes. Discourse API can stream content outcomes, returning fractional

acknowledgment comes about as they wind up accessible, with the perceived content showing up promptly while

talking. Then again, Speech API can return perceived content from sound put away in a document.

Wolfram Alpha Engine API Interface

Wolfram Alpha acquaints an on a very basic level new route with find learning and solutions not via looking

through the web, but rather by doing dynamic calculations in view of an immense accumulation of inherent

information, calculations, and techniques. The inquiry is then passed for the Wolfram Alpha Engine to figure the

appropriate response. Clients can enter inquiries in content or different other record configurations, for example,

Binary, JPEG, PDF, WAV, XML and so on. It covers 30+ information territories and orders. Acknowledges totally

freestyle input. The primary favourable position is that it creates effective outcomes and presents them with most

extreme clearness. It is a simple interface. It goes about as an information store and additionally a web crawler.

TTS Module Interface

In the wake of figuring the inquiry the outcome must be made talk, for this reason TTS module interfacing must be

made. A content to-discourse (TTS) framework changes over typical dialect content into discourse. A PC

framework utilized for this intention is known as a speech synthesizer, and can be actualized in programming or

equipment items. A content to-speech framework (or "motor") is made out of two sections: a front-end and a

back-end. The front-end has two noteworthy assignments. To start with, it changes over crude content containing

images like numbers and truncations into what might as well be called composed out words. The front-end at that

point relegates phonetic translations to each word, and partitions and denotes the content into prosodic units,

similar to expressions, statements, and sentences. Phonetic translations and prosody data together make up the

emblematic etymological portrayal that is yield by the front-end. The back-end frequently alluded to as the

synthesizer then changes over the emblematic etymological portrayal into sound.

4. EXPERIMENTAL VALIDATION AND RESULTS

The main steps involved in Personal Intelligent Assistant for Speech Recognition are,

1. Capturing the voice.

(5)

3. Searching in the Internet.

4. Processing back text to voice.

The Background steps involving are,

1. Run the Shell Script Program.

2. Run the Python Script.

The working in detail, the query is taken as the input voice by system using Microphone. The voice is recorded and

it is taken for sampling. The sample voice is converted into text and it is searched in Internet using Wi Fi or

Ethernet. Then the text is again converted back into speech and played using Microphone.

Fig.2 Final project

The code and the hardware setup has been tested. The answer is given as speech through the Headset.

5. CONCULSION AND FUTURE WORK

Speech Recognition will change the way individuals cooperated with Smart gadgets and will, ultimately, separate

the up and coming innovations. All the shrewd gadgets coming today in the market are fit for perceiving discourse.

Numerous territories can profit by this innovation. Discourse Recognition can be utilized for instinctive task of PC

based frameworks in every day life. This innovation will bring forth progressive changes in the advanced world and

turn into a rotate innovation. Inside five years, speech recognition innovation will turn out to be so unavoidable in

our every day experience that administration situations without this innovation will be viewed as inferior. Future

degree will towards to accomplish Ability to recognize subtleties of discourse and implications of words and

furthermore converse with all other devices. The remain solitary SR will be actualized soon in coming years. After

a few years the SR framework ca be executed in wearable gadgets.

REFERENCES

[1] JOE TEBELSKIS {1995}, SPEECH RECOGNITION USING NEURAL NETWORKS, School of

(6)

speech, Umeå University, Department of Philosophy and Linguistics[3]

[3] KLAUS RIES {1999}, HMM AND NEURAL NETWORK BASED SPEECH ACTDETECTION,

International Conference on Acoustics and Signal Processing (ICASSP’99)[4]

[4] B. PLANNERER {2005}, AN INTRODUCTION TO SPEECH RECOGNITION

[5] KIMBERLEE A. KEMBLE, AN INTRODUCTION TO SPEECH RECOGNITION, Voice Systems

Middleware Education, IBM.

[6] Speaker Independent Connected Speech Recognition- Fifth Generation Computer Corporation".

Fifthgen.com. Retrieved 15 June 2013.

[7] "British English definition of voice recognition". Macmillan Publishers Limited. Retrieved 21 February

2012.

AUTHOR BIOGRAPHIES

Jancy M P is a Student at Sahrdaya College of Engineering and Technology, Kerala, India. She is

pursuing her Bachelor’s degree in Computer Science and Engineering. Her area of interest is speech

recognition and data mining. E-mail: jancydundu6@gmail.com

Doniya Antony is a Student at Sahrdaya College Of Engineering and Technology, Kerala, India.

She is pursuing her Bachelor’s degree in Computer Science and Engineering. Her area of interest is

speech recognition and data mining. E-mail: doniyaantony4@gmail.com

Satheesh Kumar R is an Associate Professor in Sahrdaya College of Engineering & Technology.

He received his Master’s degree in Information Technology and MBA degree in Information

Systems. He also received his PhD degree in Information and Communication Engineering. His

research areas are Artificial Intelligence, Natural Language Processing, Network Security, Data

Mining, Web Mining, Text Mining and Opinion Mining.

Figure

Fig.1 Capturing voice query

References

Related documents

Conclusions: The findings of this study reveal that parent – child communication regarding sexual health promotion within Black families in Nova Scotia remains varied and is

S. In recent years there has been considerable progress in the area of microprocessor controlled automated test systema for the measurement of magnetic properties

2) Filtration specification for a micro-sprinkler irrigation system fed with medium quality of water : Two levels of filtration should be used; an 80 µm mesh filter at the

For contaminated soil the following genera were observed: Aspergillus, Penicillium, Rhizopus, Fusarium, Geotridum and Cladosporium, while in the artificially

Original Article Roles of insulin like growth factors and their binding proteins in the differentiation of mouse tongue myoblasts AKIRA YAMANE*,1 TOKI URUSHIYAMA2 and THOMAS G H

International Journal of Scientific Research in Computer Science, Engineering and Information Technology CSEIT195186 | Received 09 Feb 2019 | Accepted 18 Feb 2019 | January February 2019

The spectrums of the complexes formed between bupropion hydrochloride with Cu (II), Ca (II), Cd (II), Mg (II) and Zn (II) metal cations were taken in the range of 500 to 190 nm

ment discute recemment l'O.T.A.N. Par consequent, ce que nous souhaitons, Monsieur le President, ce n' est pas seulement la consolidation de !'Europe des Six, mais