Friday, 24 June 2016

GOOGLE IO 2016



Google isn't certain what to name Android N, so it's enticing the general population to submit thoughts. It reserves the privilege to pick the champ, in any case, so don't anticipate that the new Android will be called "Namey McNameface."

Android N is additionally getting upgrades on the representation and runtime execution front, security improvements (consistent overhauls and document based encryption, to be specific), and less demanding multi-tasking. An all out adaptation of Android N will be accessible to everybody later this late spring.

The main major uncover was Google Assistant, another individual AI for clients. It gives clients a chance to ask inquiries much as they would in the internet searcher, however in a Siri-like set-up.

You can request Pablo Picasso's first name, sports scores and to play a melody you've had latched onto your subconscious mind throughout the day.

Exactly where is Google Assistant going to live? In Google Home, obviously.



Pichai gave a yell out to Amazon Echo in declaring the new gadget, which is a white-and-dim Wi-Fi speaker that helps you handle ordinary assignments. It plays music and gives you a chance to control brilliant home gadgets, including Nest items. You can, obviously, ask Google Home anything you need to know, a la Google look.

Hunt is inherent, drawing on 17 years of development to "answer questions troublesome for different colleagues to handle."



You can check an eatery reservation, and, on the off chance that it needs to change, have Google Home message a companion that you'll be eating later than anticipated. You can likewise "make the most of your diversion more effortlessly than any time in recent memory," set alerts, and different dull assignments you'd preferably have an individual associate handle than yourself.

Google Home will be accessible later this fall.

Google went directly into its next declaration, another application called Allo. It incorporates Google Assistant mix, brilliant photograph acknowledgment and an emphasis on emoji. It takes advantage of neural systems and Google pursuit to tidy up your content discussions.

Allo takes advantage of your discussion history to concoct recommendations of what it supposes you need to say as a reaction. Reactions will be customized taking into account how you banter too.

It highlights an Incognito Mode, which incorporates end-to-end encryption, message termination and private warnings. On the off chance that you close down the visit it'll be erased and gone always, ideal for private discussions.

Allo will be accessible later this mid year for iOS and Android.

Google has reported Daydream, another VR stage based on Android N that will arrive this harvest time. Like the home perspective you find within Oculus Rift, Google Daydream is a holding nothing back one experience that brings amusements, applications, motion pictures and even the Google Play Store completely into a VR headset.

There was no Android VR headset to flaunt, even on Day 2 of Google IO, however Google has thought of a reference plan for different producers to work off of and additionally contend with. It's the same thought as its Nexus telephone gadgets.

Google additionally showed a little, Wii-like remote that gives movement control 15 distinctive approaches to collaborate with VR diversions as such, which means this model equipment could develop down the line.

A few Daydream-prepared gadgets will dispatch this year from any semblance of Samsung, HTC and other prevalent producers. That Huawei VR headset we tried out a month ago may have been an early review of what's in store from outsider Daydream VR outlines.

ou'll soon have the capacity to download Android applications by means of the Google Play Store on Chrome OS PCs, affirming gossipy tidbits that Google's two working frameworks would play together pleasantly at Google IO.

Out of the blue, those best Chromebook portable PCs look like genuine answers for your ordinary figuring undertakings on account of the expansion of a huge number of valuable Android applications.

Toward the begin, applications work with a trio of touchscreen-empowered portable PCs: ASUS Chromebook Flip, Acer Chromebook R11 and the most recent Chromebook Pixel. There's more cross-similarity to come, as well.

Strangely, while this feels like the starting phases of the Android being collapsed into Chrome OS, the enormous news didn't get reported in front of an audience at the Google IO keynote. Happy you read this, privilege?

Android Wear 2.0

DOTCOM BUBBLE




The dotcom bubble began without the internet, and to be sure in the first place it didn't perceive the Internet as essential. When Al Gore started discussing the "data superhighway" in the mid 1990s, in any case, the "huge end of town" - Hollywood, Silicon Valley, information transfers transporters, link organizations, and media combinations, all started contributing.

Between April 1992 and July 1993 the majority of the real US business magazines had distributed real components on new interchanges and the "Data Superhighway". It merits investigating what these magazines and highlight articles discussed. The principal thing I saw - not one of the component articles I grabbed said the Internet. It wasn't on the business skyline of this valiant new joined universe of Silicon Valley and Hollywood. They were more inspired by intelligent TV.

Business Week's July 12 1993 version had a main story "Media Mania… advanced - intelligent - interactive media - the surge is on". Time Warner's Gerard Levin discussed changing home TVs to "anything, anyplace". Electronic books and magazines were going to change the world. Intelligent TV would get to 20% of US homes by the turn of the century.

Gerard Levin was additionally in Newsweek's version of May 31 1993. The main story was a zillion dollar industry. The dotcom lack of interest to the quantity of zeroes in money related figures appears to have had its inceptions about this time. Levin was going to get his bank parities on TV. Lounge chair potatoes would have the capacity to individualize the endings of motion pictures and select camera plots for donning occasions. Clever specialists in the icebox would advise the auto to recollect that it was out of milk. (Oddly, we are as yet discussing these things 10 years after the fact).

California Business in April 1992 had Silicon Valley meeting Hollywood in a 100 billion business sector as its main story. What's more, Forbes Magazine on April 13 1992 highlighted link organizations beating the telephone organizations to wire homes for the advanced age. What's more, touted a definitive union gadget, where the TV phone and PC would converge into a solitary wise box - a telecomputer.

Anybody understanding every one of this and missing the plot of the fast approaching entry of the Internet could be totally pardoned. None of these articles gave the Internet a notice.

This helps us to understand that the Internet didn't catalyze the dotcom bubble. It was just locked on to as a vehicle when different roads for venture did not have all the earmarks of being going anyplace. The air pocket was the second California Gold Rush and advanced union before it got to be dotcom. The Dotcom madness was truly about something that didn't happen and didn't have a speck at any rate. Since large portions of the first dreams didn't look like occurrence, the landing of the World Wide Web and an alluring Internet brought on the majority of the above gatherings to change gear.

Preceding 1994, information transfers organizations were principally keen on creating more brilliant telephones, which would resemble PCs. It presumably took an additional 10 years before we began to see the kind of advancements they visualized showing up in the cell telephone enclosure.

Television and link organizations were into intuitive TV with 500 stations additionally, intuitiveness, and video on interest. Pick your own plots for games, pick you possess plots for motion pictures. Indeed, even Microsoft thought this was prone to be the principle diversion, and Microsoft turned up at link demonstrates touting new route screens for the going to-be 500 station TV set. However TV has been the love seat potato of the advanced age. It appears to be identical, to a great extent does likewise, as 10 years back.

In the event that TV is the love seat potato of the computerized age, the non-arranged PC is not a long ways behind. It truly is difficult to contend that PCs as stand alone gadgets have enhanced much since, in spite of the building proof of expanded force and usefulness. PCs stay as lousy and confounding as they were 10 years back. The last awesome advances in standalone figuring were the mouse and Windows. Dependability does not appear to have made strides. Speed for normal assignments, (for example, opening a word processor) does not seem, by all accounts, to be any quicker, albeit some additional usefulness is accessible.

The organized PC however remains as the wonders which has most influenced our lives and brought on changes. The movement of the PC from a computational to an imparting gadget is maybe the most huge change of the data innovation age in this way. The development of the Internet as a medium for associating these conveying gadgets is, I propose, the real change that happened.

Thus the net developed. For the following five years we were to be assaulted with infrequently reasonable and regularly doubtful dreams without bounds; we knew about data superhighways, web coolers and autos, learning economies, web time and web years, which were endlessly distinctive to at whatever time known some time recently, and the dotcom furor.

Not following the South Sea Island rise in the 1700s had western economies experienced anything like the dotcom monetary air pocket. All of a sudden everybody needed a slice of the profits; ordinarily canny financial specialists went insane, and mums and fathers added to the free for all. For a few, the dotcom time saw an accumulating of awesome riches. Be that as it may, overnight it vanished amid 2000 and 2001. The data age prophets of awesome things to come vanished alongside the money related benefits, and we as a whole changed in accordance with a more ordinary life, though one significantly improved by the expansive scale appropriation of the Internet in western nations.

It might take a couple of years before we know the amount of riches was lost in the dotcom time; some organizations are as yet changing in accordance with post dotcom reality. In any case, the misfortunes are positively in the billions, and with a couple of more years separation may be seen to be the main consideration in the late decrease in the US economy. However dotcom is still excessively near us to have the capacity to completely get it.

What is a Mainframe




"What is a Mainframe", and having been posed this question myself commonly I might want to propose an all the more lighting up definition. In the first place, be that as it may, some exceptionally concise anecdotal information. I first got to be keen on registering machines as a young person. In those days the second era was quickly attracting to a nearby and System/360 was going to change the figuring scene. My first programming background was in secondary school, where my class had entry to a quick IBM 7094-II (and before you ask, no, my secondary school did not have its own 7094; we were permitted constrained utilization of one of MIT's frameworks). In school I majored in math, basically on the grounds that software engineering as a noteworthy was still around 4 years later on. In any case, my first love has dependably been registering machines, and I have contributed a lifetime of study and work in this industry. I have worked with all stages aside from vector preparing based supercomputers. My most loved has dependably been, and stays right up 'til today, the centralized computer.

One may assume that it is anything but difficult to characterize a centralized server, however such is not the situation. A few definitions are broad to the point that they incorporate all registering stages. Others look to focus on some specific part of centralized server processing, (for example, the working frameworks which keep running on a centralized server) and announce that a centralized server is what runs or backings this figuring perspective. This last definition experiences two issues:

1) it is totally unenlightening; and

2) it is misdirecting. For instance, the FLEX/ES test system permits one to run OS/390, VM, and VSE/ESA on a quick Intel processor. However a great many people who have worked with both classes of machine would naturally consider the Intel PC to be the inverse of a centralized computer.

Additionally, in the level headed discussion between customer/server situated processing, and centralized computer based arrangements, the failure to unmistakably characterize the last has fetched more than one server farm its centralized computer. The "new worldview" announced that a bunching of little, restricted engineering machines, interconnected by elaborate topologies, was the influx without bounds. Lost to a nontechnical senior administration was the way that in actualizing this new computational model they were in the meantime wiping out the most effective, complete, and refined class of registering stages ever conveyed to the commercial center.

So what is a centralized computer? So as to answer this inquiry I sat down one weekend and investigated the historical backdrop of centralized computer registering, focusing on those components that are one of a kind to the centralized server world. The consequence of this exertion was the accompanying definition, which has the double points of interest of being both succinct and exact. It likewise welcomes elaboration and serves as the beginning stage for an inside and out talk of the issues it raises:

"A centralized server is a ceaselessly advancing universally useful processing stage consolidating in it compositional definition the key usefulness required by its objective applications."

Some extra remarks about this definition are all together. A standout amongst the most basic elements of the centralized computer world is the quick and obviously perpetual advancement of the product offering. From 16 general and 4 gliding point registers of System/360, to the control register increases in the mid 370s, to the entrance registers of the last 370s, to the full supplement of skimming point registers of System/390 and the full 64 bit execution offered by the z800/900 models; from 6 selector channels to 16 square multiplexing channels to 256 fast optical channels; from 142 directions to more than 500 guidelines; from genuine tending to virtual tending to virtual machines; from the basic 8 bit memory of the 360/30 through eras of advancement to the multiported, multilevel reserving, multiprocessor supporting memory of the z900, the whole equipment space of the centralized computer world has been portrayed by an unmatched, and without a doubt quickening, development.

Amid a significant part of the initial 20 years of the present day centralized server period (which started on April 7, 1964) singular models of the centralized computer line were focused by aggressive frameworks vigorously upgraded to give a predominant value/execution item inside a very much characterized specialty market. As the centralized server advanced through item revive cycles and new item declarations, the corner advantage offered by these uncommon reason contenders was underestimated, and their capacity to contend in a business sector that requested an ever more prominent universally useful ability was essentially overpowered.

The most basic characterizing component of the centralized server worldview is that the arrangements it gives are actualized fundamentally in equipment, including microcode, a methodology (in opposition to what numerous clients of different stages may envision) that is really interesting to the centralized computer world. From the early RPQs of the 360 time, to the various "helps" of the essential 370 time, to the all out compositional upgrades of the late 370 and 390 periods the centralized computer has been an equipment test bed of unmatched extension and adaptability. By method for examination, you may review that a couple of years prior Intel added about six directions to its line of Pentium processors to encourage representation preparing. Their declaration took a specific pride in taking note of this was the primary change to the PC's guideline set in the past 13 years!

A standout amongst the most striking components of centralized computer registering, when seen after some time, is the degree to which the engineering changes to oblige client prerequisites. One of the early offering purposes of System/360 was its stand-alone copying of second era frameworks. When System/370 tagged along, stand-alone imitating was supplanted by coordinated copying, a basic client necessity. Several RPQs have been made accessible throughout the years to fulfill some client necessity. Some of these arrangements were restricted time offerings; others turned into a perpetual part of the design. One of my top choices from the previous gathering was the High Accuracy Arithmetic Facility (HAAF) accessible on the IBM 4361. This centralized server, promoted as a supermini, was focused at college math and material science divisions. With establishment of the HAAF one could do coasting point math without conveying a trademark in the skimming point number. In addition, all blunders presented by part (mantissa) moving were dispensed with. This office allowed drifting point number-crunching to be broke down for precision under an extensive variety of computational conditions, a staggering ability for the math and material science clients.
HOME MATERIAL DESIGN

Android Getting Started with Material Design 



You may have known about android Material Design which was presented in Android Lollipop rendition. In Material Design parcel of new things were presented like Material Theme, new gadgets, custom shadows, vector drawables and custom activitys. On the off chance that you haven't taking a shot at Material Design yet, this article will give you a decent begin.

In this instructional exercise we are going to take in the fundamental strides of Material Design advancement i.e composing the custom subject and executing the route drawer utilizing the RecyclerView.

Experience the beneath connections to get more learning over Material Design.

> Material Design Specifications

> Creating Apps with Material Design

DOWNLOAD CODE

VIDEO DEMO



1. Downloading Android Studio

Before going further, download the Android Studio and do the essential setup as I am going to utilize Android Studio for all my instructional exercise starting now and into the foreseeable future. On the off chance that you are attempting the Android Studio surprisingly, go the diagram doc to get complete review of android studio.



2. Material Design Color Customization

Material Design gives set of properties to redo the Material Design Color subject. In any case, we utilize five essential ascribes to tweak general subject.

colorPrimaryDark – This is darkest essential shade of the application mostly applies to notice bar foundation.

colorPrimary – This is the essential shade of the application. This shading will be connected as toolbar foundation.

textColorPrimary – This is the essential shade of content. This applies to toolbar title.

windowBackground – This is the default foundation shade of the application.

navigationBarColor – This shading characterizes the foundation shade of footer route bar.

android-material-plan shading mapping

You can experience this material outline shading designs and pick the one that suits your application.




3. Making Material Design Theme

1. In Android Studio, go to File ⇒ New Project and fill all the points of interest required to make another undertaking. When it prompts to choose a default action, select Blank Activity and continue.

2. Open res ⇒ values ⇒ strings.xml and include underneath string values.

strings.xml



Material Design

Settings

Search

Open

Close

Home

Friends

Messages

<!- - route drawer thing names - >



@string/nav_item_home

@string/nav_item_friends

@string/nav_item_notifications


Messages

Friends

Home






3. Open res ⇒ values ⇒ colors.xml and include the beneath shading values. On the off chance that you don't discover colors.xml, make another asset record with the name.

colors.xml





#F50057

#C51162

#FFFFFF

#FFFFFF

#000000

#FF80AB





4. Open res ⇒ values ⇒ dimens.xml and include beneath measurements.

dimens.xml



<!- - Default screen edges, per the Android Design rules. - >

16dp

16dp

260dp






5. Open styles.xml under res ⇒ values and include beneath styles. The styles characterized in this styles.xml are regular to all the android forms. Here I am naming my topic as MyMaterialTheme.

styles.xml












6. Presently under res, make an envelope named values-v21. Inside qualities v21, make another styles.xml with the beneath styles. These styles are particular to Android Lollipop as it were.

styles.xml










7. Presently we have the fundamental Material Design styles prepared. With a specific end goal to apply the topic, open AndroidManifest.xml and alter the android:theme characteristic of tag. 

android:theme="@style/MyMaterialTheme"

So in the wake of applying the subject, your AndroidManifest.xml ought to look like underneath.

AndroidManifest.xml




package="info.androidhive.materialdesign" >


android:allowBackup="true"

android:icon="@mipmap/ic_launcher"

android:label="@string/app_name"

android:theme="@style/MyMaterialTheme" >


android:name=".activity.MainActivity"

android:label="@string/app_name" >















Presently in the event that you run the application, you can see the notice bar shading changed to the shading that we have said in our styles.

android-material-plan notice bar

3.1 Adding the Toolbar (Action Bar)






8. Make a xml document named toolbar.xml under res ⇒ format and include android.support.v7.widget.Toolbar component. This make the toolbar with particular tallness and theming.

toolbar.xml




xmlns:local="http://schemas.android.com/apk/res-auto"

android:id="@+id/toolbar"

android:layout_width="match_parent"

android:layout_height="wrap_content"

android:minHeight="?attr/actionBarSize"

android:background="?attr/colorPrimary"

local:theme="@style/ThemeOverlay.AppCompat.Dark.ActionBar"

local:popupTheme="@style/ThemeOverlay.AppCompat.Light"/>




9. Open the format record of your principle movement (activity_main.xml) and include the toolbar utilizing tag. 

activity_main.xml


xmlns:tools="http://schemas.android.com/apparatuses"

android:layout_width="match_parent"

android:layout_height="match_parent"

tools:context=".MainActivity">


android:layout_width="fill_parent"

android:layout_height="wrap_content"

android:layout_alignParentTop="true"

android:orientation="vertical">


android:id="@+id/toolbar"

layout="@layout/toolbar"/>





Run the application and check whether the toolbar showed on the screen or not.

android-material-plan toolbar

Presently we should attempt to include a toolbar title and empower the activity things.




10. Download this inquiry symbol and import it into Android Studio as an Image Asset.




11. To import the Image Asset in Android Studio, right tap on res ⇒ New ⇒ Image Asset. It will demonstrate to you a popup window to import the asset. Skim the hunt symbol that you have downloaded in the above stride, select Action Bar and Tab Icons for Asset Type and give the asset name as ic_search_action and continue.

android-studio-importing-picture resource



12. Once the symbol is foreign, open menu_main.xml situated under res ⇒ menu and include the inquiry menu thing as specified underneath.

menu_main.xml


xmlns:app="http://schemas.android.com/apk/res-auto"

xmlns:tools="http://schemas.android.com/devices"

tools:context=".MainActivity">


android:id="@+id/action_search"

android:title="@string/action_search"

android:orderInCategory="100"

android:icon="@drawable/ic_action_search"

app:showAsAction="ifRoom"/>


android:id="@+id/action_settings"

android:title="@string/action_settings"

android:orderInCategory="100"

app:showAsAction="never"/>





13. Presently open your MainActivity.java and do the beneath changes.

> Extend the action from AppCompatActivity

> Enable the toolbar by calling setSupportActionBar() by passing the toolbar object.

> Override onCreateOptionsMenu() and onOptionsItemSelected() strategies to empower toolbar activity i

Friday, 20 May 2016

A TERM PAPER ON WIRELESS SENSORS BY ME

ABSTRACT
            Advancements in sensing, microelectronics and wireless communications technologies       are paving the way for the development of a new breed of integrated wireless sensing devices. The relatively simple devices that we envision are akin to the sensory receptors of the nervous system in that they are capable of detecting changes in the environment  due to stimuli. In this paper, we examine the advancements in location estimation using  wireless sensor networks which is one of its major applications. We try to extract few techniques described by the such as self configuration of wireless sensor networks using Cramer-Rao’s method, Probability grid and Experimental analysis of RSSI, Algorithms based on RSSI sampling, ROBUST position estimation, Path loss location estimation and  Nonparametric Belief Propagation for Self-Localization. This paper presents an overview of research trends and challenges in the design and implementation of large-scale             location estimation using wireless sensor network
INTRODUCTION
                                                            A wireless sensor network (WSN) consists of spatially             distributed autonomous sensors to monitor physical or environmental conditions, such             as temperature, sound, pressure, etc. and to cooperatively pass their data through the network to a main location. The WSN is built of "nodes" – from a few to several hundreds or even thousands, where each node is connected to one (or sometimes several) sensors. Each such sensor network node has typically several parts:  a radio transceiver with an internal antenna or connection to an external antenna,   a microcontroller, an electronic circuit for interfacing with the sensors and an energy       source, usually a battery or an embedded form of energy harvesting. The WSN is built of     "nodes" – from a few to several hundreds or even thousands, where each node is    connected to one (or sometimes several) sensors. Each such sensor network node has         typically several parts: a radio transceiver with an internal antenna or connection to an          external antenna, a microcontroller, an electronic circuit for interfacing with the sensors         and an energy source, usually a battery or an embedded form of energy harvesting.
APPLICATIONS

1.     Area monitoring

                        Area monitoring is a common application of WSNs. In area monitoring, the WSN  is deployed over a region where some phenomenon is to be monitored.

2.     Environmental/Earth monitoring

                        The term Environmental Sensor Networks, has evolved to cover many  applications of WSNs to earth science research. This includes sensing volcanoes,                            oceans, glaciers, forests, etc.

3.      Water/Waste water monitoring

 Monitoring the quality and level of water includes many activities such as                                      checking the quality of underground or surface water and ensuring a country’s                               water infrastructure for the benefit of both human and animal. The area of water                          quality monitoring utilizes wireless sensor networks and many manufacturers                            have launched fresh and advanced applications for the purpose.

4.      Agriculture

                                            Using wireless sensor networks within the agricultural industry is increasingly common; using a wireless network frees the farmer from the maintenance of wiring in a difficult environment. Gravity feed water systems can be monitored  using pressure transmitters to monitor water tank levels, pumps can be controlled   using wireless I/O devices and water use can be measured and wirelessly  transmitted back to a central control center for billing. Irrigation automation  enables more efficient water use and reduces waste.

5.      Passive localization and tracking

 The application of WSN to the passive localization and tracking of non-                                         cooperative targets (i.e., people not wearing any tag) has been proposed by                                     exploiting the pervasive and low-cost nature of such technology and the                                              properties of the wireless links which are established in a meshed WSN                                               infrastructure.

6.      Passive localization and tracking

 The application of WSN to the passive localization and tracking of non-                                         cooperative targets (i.e., people not wearing any tag) has been proposed by                                     exploiting the pervasive and low-cost nature of such technology and the                                              properties of the wireless links which are established in a meshed WSN                                               infrastructure.

7.      Smart home monitoring

 Monitoring the activities performed in a smart home is achieved using wireless                               sensors embedded within everyday objects forming a WSN. State changes to                                 objects based on human manipulation are captured by the wireless sensors                                      network enabling activity-support services.
PLATFORMS
            Hardware
One major challenge in a WSN is to produce low cost and tiny sensor nodes. There are an             increasing number of small companies producing WSN hardware and the commercial             situation can be compared to home computing in the 1970s. Many of the nodes are still in  the research and development stage, particularly their software. Also inherent to sensor network adoption is the use very low power methods for data acquisition.
            Software
            Energy is the scarcest resource of WSN nodes, and it determines the lifetime of WSNs.    WSNs are meant to be deployed in large numbers in various environments, including   remote and hostile regions, where ad-hoc communications are a key component.

            Operating systems

            Operating systems for wireless sensor network nodes are typically less complex than          general-purpose operating systems. They more strongly resemble embedded systems, for                 two reasons. First, wireless sensor networks are typically deployed with a particular      application in mind, rather than as a general platform. Second, a need for low costs and   low power leads most wireless sensor nodes to have low-power microcontrollers ensuring  that mechanisms such as virtual memory are either unnecessary or too expensive to  implement.

Literature Review
          Neiyer Correal et.al. Proposed the challenges and opportunities in the wireless            sensor    networks [1]. As advancements in sensing, microelectronics and wireless   communications technologies are paving the way for the development of a new           breed   of integrated wireless sensing devices. The relatively simple devices that we envision are    akin to             the sensory receptors of the nervous system in that they are             capable of detecting    changes in the environment due to stimuli. As are their biological    counterparts, these             \RF neurons" or neuRFon devices are endowed with the ability to associate, producing     efficient sensory networks. These pervasive wireless sensor networks may potentially            have an unprecedented impact on the way we interact with our surroundings, by   providing a sensory fabric, linking cyberspace to our surrounding             environment. Many     issues must be addressed in order to bring this unconventional communication centric             vision to mainstream. This paper presented an overview of    research trends and    challenges in the design and implementation of large-scale      wireless embedded networks.

                        As the technological advances are making ultra low-power low-cost wireless          devices on       a chip feasible. In order to achieve a vision of pervasive wireless sensor         networking researchers must address many technological challenges. This paper provides an overview of key research areas in academic, government, and industry,          with a slant      toward position location. It is our position that          position location is a key             application enabler. Research into accurate location techniques,       free of infrastructure, will translate into greater ease of installation and usefulness of sensor data. Paramount to       the success of the wireless sensor network concept is achieving unprecedented end-to-            end energy efficiency across all layers of the system architecture.    Integral to achieving   this goal is the development of experimental test beds, as they are   invaluable to the   exploration of the design space and the minimization of technical risks.       As        implementations reduce in size and energy consumption, prototypes will demonstrate  compelling applications and point in new directions for further applications.




            Neal Patwari et.al. proposed Relative Location Estimation in Wireless Sensor
Networks [2]. Self-configuration in wireless sensor networks is a general class of estimation of problems which we study via the Cramer-Rao bound (CRB). Specifically, we consider sensor location estimation when sensors measure received signal strength (RSS) or time of-arrival (TOA) between themselves and neighboring sensors. A small fraction of sensors in the network have known location while the remaining locations must be estimated. We derive CRBs and maximum likelihood estimators (MLEs) under Gaussian and lognormal models for the TOA and RSS measurements, respectively. An extensive TOA and RSS measurement campaign in an indoor office area illustrates MLE performance. Finally, relative location estimation algorithms are implemented in a wireless sensor network test bed and deployed in indoor and outdoor environments. The measurements and test bed experiments demonstrate 1 m RMS location errors using TOA, and 1 m to 2 m RMS location errors using RSS.
This article has been to show the accuracy with which wireless sensor networks can estimate the relative sensor locations. The results should help researchers determine if the accuracy possible from relative location estimation can meet their application requirements. This article began by proving that location estimation variance bounds (CRB) decrease as more devices are added to the network. Next, it was shown that CRBs can be readily calculated for arbitrary numbers and geometries of devices, and several examples were presented. Sensor location estimation with approximately 1 m RMS error has been demonstrated using TOA measurements. However, despite the reputation of RSS as a coarse means to estimate range, it can nevertheless achieve an accuracy of about 1 m RMS in a test bed experiment. Fading outliers can still impair the RSS relative location system, implying the need for a robust estimator. Future experimentation is needed to verify the variance of location estimators due to the non-ergodic nature of shadowing. Analysis can quantify the effect of ‘nuisance’ channel parameters, and can be extended to consider the effects of multi-user interference on sensor location estimation.




            Radu Stoleru et.al. proposed Location Estimation Scheme for Wireless Sensor Networks using probability grid [3]. Location information is of paramount importance       for Wireless Sensor Networks (WSN). The accuracy of collected data can significantly        be affected by an imprecise positioning of the event of interest. Despite the importance of       location information, real system implementations, that do not use specialized hardware         for localization purposes, have not been successful. In this paper, they proposed a location estimation scheme that uses a probabilistic approach for estimating the             location of anode in a sensor network. Our localization scheme makes use of additional knowledge of topology deployment. We assume a sensor network is deployed in a controlled manner, where the goal of the deployment is to form a grid topology. We          evaluate our localization scheme through simulations, showing localization errors as low      as 3% of radio range. We outperform similar localization schemes by obtaining 50%    less error in localization, when compared to them. We also evaluate our localization solution and the DV-Hop scheme in a real implementation, obtaining an average error in         location of 79% of radio range, outperforming DV-Hop by approximately 40%. We          analyze the significant differences in        performance between simulations and a real implementation and stress the importance of         further evaluations of real implementations.   The result is an effective and realistic protocol that works in an actual   implementation, under certain assumptions, because it exploits deployment information.
                        In this paper they presented a localization scheme, called the Probability Grid that can be used in WSN which have been deployed in a grid topology. This scheme was         inspired by a similar solution, called DV-Hop, from Rutgers, which use shop count,    from anchors to sensor nodes, as a measure of distance to known locations. The DV-Hop           scheme is more general than this solution, since it does not use the knowledge about             deployment topology. However, this generality does not help, since the errors are high in practical deployments. Probability Grid is more resistant than DV-Hop to situations when   the hop count is not a very accurate measure of the distance between two points. This is           due to the fact that in real deployments, the radio range is not circular. Previous research has shown the presence of long links, backward links or stragglers and a significant       deviation from a circular radio pattern. Using a probabilistic approach, the Probability       Grid considers the hop count for a particular distance to be a discrete random variable   that has a Poisson distribution. Our solution is completely distributed and does not require special infrastructure. It only requires, as similar solutions do, to have a small   percentage of the nodes, called anchors, aware of their position. They planned to extend             their work in the following directions: a) empirically obtain a real distribution of hop         counts for different distances. In simulations and limited experimental verification, a Poisson distribution proved satisfactory, however, from an empirically obtained          distribution, we expect higher accuracy in a real system deployment; b) each node,       through beaconing,     can acquire additional information about its neighbors positions.      This can serve as a reinforcement of accuracy of its own location computation and   eliminate cases where multiple nodes localize themselves to the same position in the grid; c) employ local flooding from anchors to the adjacent nodes. They observed that the           contribution made to   the precision in position, made by anchors further away is much             smaller than the contribution made by the anchors closer to the node; d) investigate the     possibility of             regionally centralizing the hop count information, in an attempt to better     compute the    optimal positioning of nodes in the grid, in order to decrease the total     degree of uncertainty/entropy. Entropy based approach for target localization is    proposed by Wang et al. Their entropy-based sensor selection heuristic objective is to             reduce the entropy of the target location distribution. Similarly, they plan on using the       centralized approach in order to reduce the entropy of the sensor location distribution.


            Mohit Saxena et.al. proposed Experimental Analysis of RSSI-based Location
Estimation in Wireless Sensor Networks [4]. With a widespread increase in the number of mobile wireless systems and applications, the need for location aware services has risen at a very high pace in the last few years. Research has been done for the development of new models for location aware systems, but most of it has primarily used the support of 802.11 wireless networks. Less work has been done towards an exhaustive error analysis of the underlying theories and models, especially in an indoor environment using a wireless sensor network. We present a thorough analysis of the Radio Signal Strength (RSS) model for distance estimation in wireless sensor networks through an empirical quantification of error metrics. Further on the basis of this experimental analysis, we implement a k - nearest signal space neighbor match algorithm for location estimation, and evaluate some crucial control parameters using which this technique can be adapted to different cases and scenarios, to achieve finer and more precise location estimates.
            In this work, their major objective has been to quantify how good and accurate is the RSSI model in a wireless sensor network to estimate the location of a cooperative target. They have classified their observations in two broad categories, the first ones are based on a calibration based analysis and the second ones are based on a full - fledged scheme for location estimation, the k - nearest signal space neighbor match algorithm. Our results are encouraging and we are able to achieve an accuracy of nearly 1.1 meters with 90% probability in indoor environment. In the first set of results, they quantified the relationship between relative error and actual distance which we empirically prove to be multiplicative. Once we have a good quantification of the signal strength model, we implemented a location estimation scheme on this basis. The first relationship which comes to surface is the variation of accuracy with changes in the control parameter k. Next they investigated the impact of variation in mote orientations on the accuracy of location estimation. Appropriate choice of k within the OCV ranges proved to give more accurate results. Choosing maximal signal strength fingerprints while building the offline signal space, makes the location estimation more independent of user orientation. They also observe that the performance of our system improves as the size of the radio map is enhanced by increasing the number of grid points - N. One of the extensions to this system built using a WSN could be to analyze how the accuracy levels vary as we increase the number of targets being tracked at the same time from one to more. As they increase the number of motes being tracked, the number of packets being sampled at the base station will increase manifold, which can in turn result in the degradation of sensitivity of the location estimates of the objects. they would also like to investigate the performance of our system under phenomena such as shadowing and signal contention between different motes and interference with other low-power wireless devices, which work on the same frequency channels.

Javad Rezazadeh et.al. proposed Fundamental Metrics for Wireless Sensor Networks localization [5]. Localization in wireless sensor networks (WSNs) is a broad topic that has received considerable attention from the research community. The approaches suggested to estimate location are implemented with different concepts, functionalities, scopes and technologies. This paper introduces a methodological approach to the evaluation of localization algorithms and contains a discussion of evaluation criteria and performance metrics followed by statistical empirical simulation models and metrics that affect the performance of the algorithms and hence their assessment. The major contribution of this paper is to analyze and identify relevant metrics to compare different approaches on the evaluation of localization schemes.
                        In this paper they presented, different methods to implement localization and the main metrics that a system designer has to take into account to understand and value the          different location-sensing systems. In the simulation results of various localization        schemes, where the accuracy was examined through the trade-offs between accuracy and             measurement performance, percentage of anchors, deployment of anchors, density of non      anchors, etc. Besides randomly generated networks, a typical deployment of nodes is the grid of non-anchor nodes within a particular area. The localization accuracy of a solution          is usually quantified using the average Euclidean distance between the estimated     locations and the true locations normalized to the radio range or other system metrics. For     mobility-assisted localization, the effect of node density is not as important as in static             localization scenarios. In addition, communication computation cost may not be of same   importance to the off-line simulations as to the real implementations.


            Charalampos Papamanthou et.al. proposed Algorithms for Location Estimation
            Based on RSSI sampling [6].In this paper, they re-examine the RSSI measurement          mode for location estimation and provide the first detailed formulation of        the probability             distribution of the position of a sensor node [5]. They also show how to use this        probabilistic model to efficiently compute a good     estimation of the position of the sensor node by sampling multiple readings from the beacons (where we do not merely      use the mean of the samples) and then minimizing a function with an acceptable           computational             effort. The results of the simulation of our           method in TOSSIM indicate that the location of the sensor node can be computed in a small amount of time and that            the quality of the solution is competitive with previous approaches.
                        In this paper, they have analyzed the RSSI model for location estimation in            sensor networks. Given a normal distribution for the error in dBm, they computed the           correct probability distribution of the sensor’s location and then they adopted this probability distribution in a theoretical analysis of sampling the measurements for        location estimation. They finally gave a simple algorithm that can be executed on sensor             nodes; its complexity, for a constant number of beacons, is proportional to the size of the sample. Location estimation in sensor networks presents several trade-offs. If higher accuracy is desired, one has to deploy more beacons or use more samples. Using a large   number of beacons and samples causes significant energy consumption. The energy-     optimal case    occurs when only three beacons are deployed and an estimation of the        actual point is          based on the probability distribution computed by taking into          consideration only one measurement. This solution, however, gives unacceptable errors.             Additionally,   performing computations with the exact probability distribution is        unrealistic, since it       involves complex formulas. Hence, they were to depend on few             measurements, off-line computed data must be stored as tables within the sensor, which    immediately creates a          storage problem. However, one can use more samples, thus increasing energy consumption.


            Loukas Lazos et.al. proposed Robust Position Estimation in Wireless Sensor    Networks [7].They  addressed the problem of secure location determination, known as Secure Localization, and the problem of verifying the location claim of a node, known as       Location Verification, in Wireless Sensor Networks (WSN). They proposed a robust           positioning system they call ROPE that allows sensors to determine their location             without any centralized computation. In addition, ROPE provides a location verification mechanism that verifies the location claims      of the sensors before data collection. They    show that ROPE bounds the ability of an             attacker to spoof sensors locations, with        relatively low density deployment of reference points. They confirmed the robustness of   ROPE against attacks analytically and via simulations.
                        They studied the problem of secure position determination and location      verification in wireless sensor networks. We proposed a sensor initiated localization            algorithm called Robust Position Estimation (ROPE), that achieves robust sensor localization and verification of sensor location claims even in the presence of malicious         adversaries. Compared to previously proposed schemes, ROPE allows sensors to estimate             their own location without the assistance of a central authority, while being resistant to     severe types of attacks such as the wormhole attack, node impersonation and jamming of   transmissions. We introduced a new metric called Maximum Spoofing Impact (MSI) for evaluating the impact of possible attacks, and showed that ROPE limits the MSI even for           low densities of reference points.


            Guoqiang Mao et.al. proposed Path Loss Exponent Estimation for Wireless Sensor
            Network Localization [8].The wireless received signal strength (RSS) based         localization techniques have attracted significant research interest for their simplicity.         The RSS based localization techniques can be divided into two categories: the distance       estimation based and the RSS profiling based techniques. The path loss exponent (PLE)           is a key parameter in the distance estimation based localization algorithms, where   distance is estimated from the RSS. The PLE measures the rate at which the RSS        decreases with distance, and its value depends on the specific propagation environment.    Existing techniques on PLE estimation rely on both RSS measurements and distance    measurements in the same environment to calibrate the PLE. However distance             measurements can be difficult and expensive to obtain in some environments. In this         paper they proposed several techniques for online calibration of the PLE in wireless     sensor networks without relying on distance measurements. They demonstrated that it is            possible to estimate the PLE using only power measurements and the geometric        constraints associated with planarity in a wireless sensor network. This may have a significant impact on distance-based wireless sensor network localization.
                        In this paper, they presented some techniques for online calibration of path loss      exponent in wireless sensor networks without relying on distance measurements.           Specifically, techniques were proposed which are based on different assumptions about
            knowledge of distance information. The first technique assumes that the probability          distribution of distance between neighboring sensors is known. Then an algorithm similar to the quantile-quantile plot was proposed, which can estimate the path loss exponent            accurately using a small number of received power measurements. However this assumption of knowing the distance distribution can be unrealistic in some applications.    This has motivated us to find a more generic technique without using any distance             information. Then they presented a technique based on the Cayley-Menger determinant,   which estimates the path loss exponent using only power measurements and the       geometric constraints associated with planarity in a wireless sensor network. The           technique can give an accurate estimate of alpha when there is no noise in power measurements, but it has a large bias in the presence of noise. A pattern matching   technique approximately correcting the bias is proposed based on the empirical           observation that the relationship between E(^alpha), sigma db and alpha is independent of            the distribution of the vertices of various quadrilaterals and the shape of the area in which    vertices of the quadrilaterals are located. They also presented an improvement of the             earlier technique using data fusion. The proposed algorithms may have significant             impact on distance-based wireless sensor network localization, where distance is         estimated from the received signal strength measurements. In this paper, we observed the         empirical law that the relationship between E(^alpha), sigma dB and alpha is independent    of the distribution of the vertices of the quadrilaterals and is also independent of the   shape of the area in which the vertices of the quadrilaterals are located. It is desirable to                obtain an analytical expression of the relationship between E(^alpha), sigma dB and alpha. This is the direction of our future research. Furthermore, the proposed algorithm          relies on the log-normal propagation   model in Eq. 1 and Eq. 2 in the sense that the    maximum likelihood estimator shown in Eq. 11 may have a different form when the             received signal strength has a different model. Although the log-normal propagation          model is a popular model for wireless networks, there are environments in which the log-         normal propagation model is not the best model. In that case, a technique needs to be developed to select the best model and choose the best estimator for distance to replace      Eq. 11 accordingly. Therefore how to develop an algorithm for environments in which      the log-normal propagation model does not apply is also a future research topic.


            Om Prakash Sahu et.al. proposed Practical Solution for Location Estimation in          Manually Deployed Wireless Sensor Networks [9]. This paper addresses the existing           research and adds another aspect of functionality by incorporating pertinent sensor nodes       to provide a dynamic location discovery and estimation. The software used provides an           easy graphical user interface to visualize a particular location in accordance with     geographical latitude   and longitude. A simple real time location estimation technique is    worked out for wireless sensor networks based on manual deployment of sensors. The        proposed scheme finds more efficient solutions with less quantity of sensors as compared   to existing deployment schemes. The set up is evaluated exclusively in real environments         using IRIS sensor nodes supported by a global positioning system module to provide            visualization of an outdoor location. The results are offered by Google Earth application.
                        The technique to deploy sensor nodes manually is currently used in several             projects, and there are scenarios of real system deployments, where manual deployment           is the only solution. Results show that the deployed nodes estimate their relative latitude          and longitude positions for a reference point. The average localization error is mainly due           to the limitations of the devices used, because for location information, the nodes   completely rely on the global positioning system and localizing themselves in the middle   of their proximate placements or reference points.

                        In future, performance evaluation and scalability will be done through simulation. The experience from the current deployment of the sensors can be used further to address the aerial deployment. Considering the rate, altitude, and trajectory of sensor nodes the        actual location information at the time of initial deployment, can also be obtained using             our solution, giving a starting point towards a better and precise localization scheme. The      authors are working towards making it more common and adaptable for other user  communities also.


            Alexander T.Ihler et al. proposed Nonparametric Belief Propagation for Self- Localization of Sensor Networks [10]. Automatic self-localization is a critical need for     the effective use of ad hoc sensor networks in military or civilian applications. In general,    self-localization involves the combination of absolute location information (e.g., from a         global positioning system) with relative calibration information (e.g., distance             measurements between sensors) over regions of the network. Furthermore, it is generally   desirable to distribute the computational burden across the network and minimize the   amount of inter sensor communication. They demonstrated that the information used for      sensor localization is fundamentally local with regard to the network topology and use      this observation to reformulate the problem within a graphical model framework. We             then     present and demonstrate the utility of nonparametric belief propagation (NBP), a   recent generalization of particle filtering, for both estimating sensor locations and             representing location uncertainties. NBP has the advantage that it is easily implemented in a distributed fashion, admits a wide variety of statistical models, and can represent   multimodal uncertainty. Using simulations of small to moderately sized sensor networks,        we show that NBP may be made robust to outlier measurement errors by a simple model             augmentation, and that judicious message construction can result in better estimates.         Furthermore, we provide an analysis of NBP’s communications requirements, showing         that typically only a few messages per sensor are required, and that even low bit-rate           approximations of these messages can be used with little or no performance impact.
                        They proposed a novel approach to sensor localization, applying a graphical            model framework and using a nonparametric message-passing algorithm to solve the   ensuing inference problem. The methodology has a number of advantages. First, it is     easily   distributed (exploiting local computation and communications between nearby             sensors), potentially reducing the amount of communications required. Second, it computes and             makes use of estimates of the uncertainty, which may subsequently be        used to determine the reliability of each sensor’s location estimate. The estimates easily         accommodate complex, multimodal uncertainty. Third, it is straightforward to            incorporate additional sources of information, such as a model of the probability of             obtaining a distance measurement between sensor pairs. Finally, in contrast to other           methods, it is easily extensible to non-Gaussian noise models, which may be used to        model and increase robustness to measurement outliers. In empirical simulations, NBP’s          performance is comparable to the centralized MAP estimate, while additionally    representing the inherent uncertainties. They have also shown how modifications to the       NBP algorithm can result in    improved performance. The NBP framework easily accommodates an outlier process model, increasing the method’s robustness to a few             large errors in distance measurements for little to no computation and communication   overhead. Also, carefully chosen proposal distributions can result in improved small-   sample performance, reducing the computational costs associated with calibration.             Finally, appropriate message schedules require very few message transmissions, and       reduced-complexity representations may be             applied to lessen the cost of each message     transmission with little or no impact on the    final solution. There remain many open         directions for continued research. First, other message-passing inference algorithms (e.g.,    max-product) might improve performance if adapted to high-dimensional non-Gaussian    problems. Also, alternative graphical model representations may bear investigating; it           may be possible to retain fewer edges, or improve the accuracy of BP by clustering      nodes. Given its promising initial performance and many possible avenues of       improvement, NBP appears to provide a useful tool for estimating unknown sensor            locations in large ad hoc networks.


DISCUSSION AND CONCLUSION

            Technological advances are making ultra low-power, low-cost wireless devices on a chip   feasible. In order to achieve a vision of pervasive wireless sensor networking researchers        must address many technological challenges. This paper provides an overview of key re-
            Search areas in academia, government, and industry towards position location. This           article began by proving that location estimation variance bounds (CRB) decrease as            more devices are added to the network. Next, it was shown that CRBs can be readily            calculated for arbitrary numbers and geometries of devices, and several examples were             presented. Sensor location estimation with approximately 1 m RMS error has been             demonstrated using TOA measurements. In the next paper Using a probabilistic     approach, the Probability Grid considers the hop count for a particular distance to be a
            Discrete random variable, that has a Poisson distribution. Another techniques objective     has been to quantify how good and accurate is the RSSI model in a wireless sensor      network to estimate the location of a cooperative target. In this article the next one was    different methods to implement localization and the main metrics that a system designer            has to take into account to understand and value the different location-sensing systems. In      the next part we studied the problem of secure position determination and location
            Verification in wireless sensor networks. The next part presented some techniques for        online calibration of path loss exponent in wireless sensor networks without relying on           distance measurements. Specifically, techniques were proposed which are based on           different assumptions about knowledge of distance information. In the last part a novel approach to sensor localization, applying a graphical model framework and using a             nonparametric message-passing algorithm to solve the ensuing inference problem.

REFERENCES

[1]         Neiyer Correal and Neal Patwari “Wireless Sensor Networks: Challenges and       Opportunities”, Florida Communications Research Labs Motorola Labs 8000 West Sunrise Blvd, Rm 2141 Plantation, FL 33322.
[2]        Neal Patwari, Alfred O. Hero III, Matt Perkins, Neiyer S. Correal and Robert J. O’Dea     “Relative Location Estimation in Wireless Sensor Networks”.
[3]        Radu Stoleru and John A. Stankovic “Probability Grid: A Location Estimation Scheme     for Wireless Sensor Networks”, Department of Computer Science University of Virginia
            Charlottesville, VA 22903 {stoleru, stankovic@cs.virginia.edu}
[4]        Mohit Saxena, Puneet Gupta and Bijendra Nath Jain “Experimental Analysis of RSSI-      based Location Estimation in Wireless Sensor Networks”.
[5]        Javad Rezazadeh, Marjan Moradi andAbdul Samad Ismail “Fundamental Metrics for        Wireless Sensor Networks localization”, International Journal of Electrical and         Computer Engineering (IJECE) Vol.2, No.4, August 2012, pp. 452~455 ISSN: 2088-        8708.
[6]        Charalampos Papamanthou, Franco P. Preparata, and Roberto Tamassia “Algorithms for   Location Estimation Based on RSSI Sampling”.
[7]        Loukas Lazos, Radha Poovendran and Srdjan Capkun “Rope: Robust Position      Estimation in Wireless Sensor Networks”.
[8]        Guoqiang Mao, Brian D.O. Anderson and Barıs¸ Fidan “Path Loss Exponent Estimation   for Wireless Sensor Network Localization”.

[9]        Om Prakash Sahu, Tarun Dubey A Practical Solution for Location Estimation in   Manually Deployed Wireless Sensor Networks”, Wireless Sensor Network, 2011, 3, 117-           120 doi:10.4236/wsn.2011.34013 Published Online April 2011 (http://www.SciRP.org/journal/wsn).
[10]      Alexander T. Ihler, John W. Fisher, III, Randolph L. Moses, and Alan S. Willsky Nonparametric Belief Propagation for Self-Localization of Sensor Networks”, Ieee           Journal On Selected Areas In Communications, Vol. 23, No. 4, April 2005 809.
[11]      UCLA. Wireless Integrated Network of Sensors (WINS).    http:nnwww.janet.ucla.edunWINSn.
[12]      J. Werb and C. Lanzl, “Designing a positioning system for finding things and people         indoors,” IEEE Spectrum, vol. 35, no. 9, pp. 71–78, Sept. 1998.
[13]      http://www.xbow.com/Products/Product_pdf_files/Wireless_pdf/ 6020-0042         05_A_MICA2.pdf.
[15]      I. F. Akyildiz, W. Su, Y. Sankarasubramaniam, and E. Cayirci, “A survey on sensor          networks” IEEE Communications Magazine, vol. 40, no. 8, pp. 102–114, 2002.
[16]      Debnath, Ashmita; Singaravelu, Pradheepkumar; Verma, Shekhar (19 December 2012).     "Efficient spatial privacy preserving scheme for sensor network". Central European    Journal of Engineering 3(1): 1–10. Doi:10.2478/s13531-012-0048-7
[17]      D. Culler, D. Estrin, and M. Srivastava, “Overview of Sensor Networks,” IEEE     Computer, Vol. 37, No. 8, pp. 41-49,2004.
[18]      Crossbow Technology, Inc. MPR/MIB Mote Hardware Users Manual, 2006. The user       manual is retrieved from http://www.xbow.com/Support/Support_pdf_files/MPRM
            IB_Series_Users_ManuaL