Showing posts with label Tech News. Show all posts

Govt. to provide Aakash 2 (Tablet) in just 1,500 INR

The Telecom and IT Minister, Kapil Sibal said that the Government ready to provide Aakash Tablet to the peoples at a cost of Rs.1,500/- and for the students for Rs.1,130. Sibal in New delhi said there that he requested the Director General of C-DAC to help the students to realize their dream of seeing carry Aakash tablet in their hands.

Launched on November 11th, 2012 by the President of India, Pranab Mukherjee, the second version of Aakash is designed, developed and manufactured by DataWind based its proposed improvements to Aakash 1. The product is currently being supplied to IIT Bombay, who in turn is tasked with application development, content integration, field testing and deployment. The reviews for Aakash 2 have been very strong globally and on November 28th, UN Secretary General Ban Ki Moon launched it at the United Nations in New York.
The commercial version of Aakash 2 is offered in the Indian market as the UbiSlate 7Ci.
We are providing a comparison between Aakash1 and Aakash2 features :

For more latest tech news and updates keep visiting us.
Please comment for any questions or suggestions.

George Bush's family emails, photos hacked

A hacker apparently accessed private photos and emails sent between members of the Bush Family, including both former presidents, and a spokesman for George HW Bush said a criminal investigation is under way. George Bush's family emails, photos hacked

George Bush's family emails, photos hacked

The Smoking Gun website said the hacker, who went by the online moniker "Guccifer," gained access to emails, photos, private telephone numbers and addresses of Bush family members and friends.
The website displayed photos it said came from the hacker, including one that purported to show the elder Bush during his recent near-two-month stay in a Houston hospital where the 88-year-old was treated for complications arising from a bronchial infection.

The authenticity of the photos and other details on the website could not immediately be confirmed. A spokesman for former president George HW Bush declined to comment on the reports.

"There's a criminal investigation and, as such, there's nothing else we can say," Jim McGrath told The Associated Press Friday.

The FBI in Houston, where Bush lives, was similarly tight-lipped.

"We do not confirm or deny the existence of any investigation," Houston FBI spokeswoman Shauna Dunlap said.

Freddy Ford, a spokesman for former president George W Bush, who has a home in Dallas, also said Friday he "won't be commenting."

The word "Guccifer" is plastered across the photos published on the Smoking Gun website in translucent, neon blue print. The website quotes "Guccifer" as describing himself as a veteran hacker who has long been in the government's sights.

Hangout with Google+ Hangout.

Google + Hangout

Connect with up to nine people at once from your computer or the Google+ mobile app (Android, iPhone). You can doodle with your kids when you’re away from home, play poker with your buddies or share the latest news and see everyone’s reactions.

Google + Hangout

"Hangouts" are places used to facilitate group video chat (with a maximum of 10 people participating in a single Hangout at any point in time). Only Google+ users can join the "Hangout" if they happen to possess the unique URL of the Hangout. On August 18, 2011 Google added a new addition to "Hangouts" - clicking on the Share button under any YouTube video reveals an icon that suggests watching the video with friends in a Google+ hangout. Google decided to limits Hangouts On Air features in few countries. Users cannot start a Hangout On Air from China, Thailand, and Vietnam.

Mobile Hangouts currently supports Android 2.3+ devices with front-facing cameras are available since September 20, 2011. As of July 10, 2012 Google+ users on iOS are able to use Hangouts on iPhone and iPad.

Hangouts On-Air gives users the ability to create instant webcasts over Google+. The broadcasts can also be recorded for later retrieval. This feature, announced on September 20, 2011, is currently limited to some videocast personalities, but the announcement indicates that it will be opened up. The first publicly broadcasted Hangout was with The Black Eyed Peas’ on the night of September 21, 2011. The feature became available at a large scale on May 7, 2012. Currently, the feature isn't available to users under age 18.

You can also Broadcast your live Hangout On Air on Google+, your YouTube channel and website, and share the public recording once you're done. You can stream a conference keynote, host a worldwide concert or moderate a panel discussion with international experts.

Keep visiting us for more latest tech news.
Please comment if you have questions or suggestions.

New iOS Facebook App Gives you free calling....!!

You all have used Voice call over Google talk but now it's also available on your favorate facebook App for iOS(Compatible with iPhone, iPod touch, and iPad. Requires iOS 4.3 or later) in Latest version 2.1.1.

The new function, which is rolling out to the Facebook Messenger iOS app, requires Messenger users to open a conversation with another iPhone owner, tap the "i" button in the top right corner, and press "Free Call."

The recipient will then receive a pop-up notification that says, for example, "Iliyas Mansuree is calling."

What's New in Version 2.1.1
- Send a quick voice message when you have more to say
- Call friends for free right from Messenger*

*Free calling uses your existing data plan, and will be rolling out over the next few weeks.

New iOS Facebook App Gives you free calling
New iOS Facebook App Gives you free calling
New iOS Facebook App Gives you free calling
Call Kate Freeman

Like and share the post on Facebook, Twitter and Google + with your friends and followers.

DRDO and PMO Websites Hacked By Algerian hackers !

DRDO and PMO Websites Hacked By Algerian hackers !

DRDO and PMO Websites Hacked !
DRDO and PMO Websites Hacked !

Algerian hackers carried out a successful cyber attack on a government server which hosts websites of extremely sensitive organisations and defaced websites operated by the Defence Research and Development Organisation (DRDO) and the Prime Minister’s Office (PMO). 

The most sensitive website that came under attack was the one operated by the Recruitment and Assessment Centre (RAC) of the DRDO (

The RAC recruits scientists for various DRDO laboratories and also assesses the suitability of DRDO scientists for promotions. The other most important website was the one hosted by the Advisor to the Prime Minister on Public Information, Infrastructure and Innovations (

The other victim websites were: West Bengal police (; Directorate of Estates of Ministry of Urban Development (, Biotechnology Industry research Assistance Council (, UT of Diu ( and’.

A Hyderabad-based cyber security expert Kusumba S. told ‘The Hindu’ that the hacker had apparently accessed the host server and modified the root files of respective websites. “This could be dangerous because the hacker could have stolen data as he had complete access to root files,” he felt.

The Hindu : News / National : DRDO website hacked

Windows 8 Released

After so much wait finally Microsoft released Windows 8. Microsoft is offering Windows 8 upgrades priced at just $39.99USD/£24.99/1999.00 INR for anyone using an existing product. The offer will be open until January 31 and after that the OS will cost a higher price.

Windows 8 is released in 3 Versions:

# For Intel based PCs: Windows 8 and Windows 8 PRO
# For ARM based tablets: Windows 8 RT

More about Windows 8:

Cool Home Screen:

Vibrant and beautiful, the Start screen is the first thing you'll see. Each tile on the Start screen is connected to a person, app, website, playlist, or whatever else is important to you. Tiles light up with the latest info, so you're instantly up to date. In one glance, you'll see that photo you were just tagged in, tomorrow’s weather and messages from your friends.


In Windows, even the little things are all about you. Like the screen that appears when your PC is locked—it can be any picture you choose. And you'll see quick notifications on your lock screen before you sign in to your PC or open a single app.
With picture password, you can sign in by tracing a pattern on a picture of your choice instead of typing a password. Zigzag over your dog’s face? Three taps on your child’s nose? The choice is yours.
Cool Apps :
Built-in apps like People, IE10, Mail, Photos, and Messaging power you through all your essential tasks and work together for a single, streamlined experience. And there’s a world of new apps to discover in the Windows Store.

Something you need to add or you will miss:
In Windows 8 not all your current Apps are supported and you also need to add more apps to use your hardware like:
Install an app to play DVDs > you may need to install an app to play DVDs in Windows 8.
Sidebar gadgets aren't supported in Windows 8 > you won’t be able to use the sidebar gadgets that are installed on your PC in Windows 8.
 Internet TV for Windows Media Center > Not compatible.

Upgrade to Windows 8 :

Before upgrading to Windows 8, Microsoft recommends that you run the Windows 8 Upgrade Assistant. It scans your current PC to see if it is ready for Windows 8 and then provides a compatibility report and optional steps for you to buy, download, and install Windows 8.
# You can run Upgrade Assistant without purchasing or installing Windows 8.
# To check if your peripheral devices like printers or monitors will work with Windows 8, be sure they're plugged in and connected to your PC before you run Upgrade Assistant.
# Use following link to download Windows 8 Upgrade assistant: 

Please share it if you like and add your comment to make it better.. :)

Application stealing your DATA

Warning from "Symantec Intelligence Report September 2012"..!!

          The world of Android applications is truly a buzzing hive of activity these days. As a result, more and more scammers jump on this highly productive bandwagon, and the types of attacks and scams get more creative—some are so incredible they defy belief.

The effectiveness of these types of applications varies from the useful to the negligible, so a little research is required to determine this. Unfortunately there are also malicious applications, such as “Battery Long” (Android.Ackposts1), that appear to help with the battery life, but simply steal information from the compromised device.

Breaking through the boundaries of credibility are a bunch of applications that will supposedly turn your phone screen into a solar charger. Even though this is completely false, there are a number of “legitimate” applications out there making this claim. Many operate by using the cameras to measure the ambient light levels to move an onscreen dial, indicating the “charge rate” for increased accuracy. These are joke applications at best, in some cases even including small print on the application description page denying it has the ability to actually charge the phone.
Beyond the fun that can be had playing practical jokes, there is good reason to avoid such applications altogether. Take the following iteration of Android.Sumzand2 for example.

The application claims to be able to convert the screen on your device into a solar panel and use it to charge the battery, if exposed to sunlight. However, there are some unstated capabilities within this application that you need to watch out for—Android.Sumzand also happens to steal contact data from your phone.

Until real solar panels are actually installed on phones, it’s best to just continue charging your phone the old-fashioned way: plugging it in to a wall socked or USB port. Besides that, be careful what you download and install from application marketplaces. If an application requests permissions that seem out of the ordinary for what it is supposed to do, then don’t install it.

NOKIA Providing your phone "a wireless charging pillow"

Nokia recently announced its latest smartphone  Lumia 920, which has wireless charging feature.
As the name implies, wireless charging is a cordless battery refill. No more cables need to be connected from the power source to the smartphone.
To recharge the battery, Lumia 920 users simply put the smartphone on a bearing charger . And the charger is of course still connected by cable to power source.
Nokia has collaborated with Fatboy to provide pads dock.

As the new Lumia 820 and Lumia 920 support Qi wireless charging, Nokia brings three new wireless charing accessories, the Fatboy Wireless Charging Pillow (DT-901), Wireless Charging Plate (DT-910) and the Wireless Charging Stand (DT-900). All of them support Qi wireless charging technology, meaning you can simply place your Lumia phone on them to start recharging.

Qi Technology :

The name "Qi" comes from the "qi" concept of energy flow from Chinese medicine, and is pronounced "chee." Qi works by generating an electromagnetic field between a charger and a device through magnetic induction. Qi chargers have a built-in transmitting coil that communicates with Qi devices using a specific electromagnetic frequency, so any Qi device will work with any Qi charger.

Under the Qi specification, "low power" for wireless transfer means a draw of 0 to 5 W. Systems that fall within the scope of this standard are those that use inductive coupling between two planar coils to transfer power from the power transmitter to the power receiver. The distance between the two coils is typically 5 mm. Regulation of the output voltage is provided by a digital control loop where the power receiver communicates with the power transmitter and requests more or less power. Communication is unidirectional from the power receiver to the power transmitter via backscatter modulation. In backscatter modulation, the power-receiver coil is loaded, changing the current draw at the power transmitter. These current changes are monitored and demodulated into the information required for the two devices to work together.

Post your comments !!
Keep Visiting for latest updates... :)


Google has flipped the switch on its free turn-by-turn voice navigation service and traffic data in India. Any Android user can now use Google Maps Navigation to select best possible route to their destination, however the live traffic data is also available on desktop maps and mobile maps from the company.


Google Now uses card format to show this information and one of the several cards is traffic card, which gives you traffic conditions and alternate routes before you leave for work or home. Google Now also puts traffic to your next likely destination at your fingertips. But until now, the traffic cards weren't of much use in India as there was no traffic data, which has now changed.

Live traffic updates within maps for selected cities as Delhi, Mumbai, Bangalore, Chennai, Hyderabad, Pune but, surprisingly, not Kolkata. Red is the color indicating heavy traffic, while green indicates free flow, and yellow is somewhere in between.

Started in first week of September 2012, if you have a Jelly Bean running smartphone with Google Now, you will be able to see estimated time to your destination (home, office or any other location you search on Google) with traffic condition details. So, if you see that a particular road is seeing high-traffic, you can take an alternate route to your office and reach on time.

Can Science Make Facebook More Compassionate?

Facebook is confronting cyberbullying and online conflict. Can a team of researchers help boost kindness among the site's 900 million users?
Eighteen months ago, Arturo Bejar and some colleagues at Facebook were reviewing photos on the site that users had flagged as inappropriate. They were surprised by the offending content—because it seemed so benign.
“People hugging each other, smiling for the camera, people making goofy faces—I mean, you could look at the photographs and you couldn’t tell at all that there was something to make somebody upset,” says Bejar, a director of engineering at the social networking site.

Then, while studying a photo, one of his colleagues realized something: The person who reported the photo was actually in the photo, and the person who posted the photo was their friend.
As the team scrolled through the images, they noticed that was true in the vast majority of cases: Most of the issues involved people who knew each other but apparently didn’t know how to resolve a problem between them.
Someone would be bothered by a photo of an ex-boyfriend or ex-girlfriend, for instance, or would be upset because they were excluded from a photo that showed a friend’s “besties.” Often people didn’t like that their kids were in a photo a relative had uploaded. And sometimes they just didn’t like the way they looked.
Facebook didn’t have ways to identify or analyze these problems, let alone resolve them. And that made Bejar and his colleagues feel like they weren’t adequately serving the Facebook community—a concern amplified by the site’s exponential growth and worries about cyberbullying among its youngest users.
“When you want to support a community of a billion people,” says Bejar, “you want to make sure that those connections over time are good and positive and real.”
A daunting mission, but it’s one that Bejar has been leading at Facebook, in collaboration with a team of researchers from Yale University and UC Berkeley, including scientists from the Greater Good Science Center. Together, they’re drawing on insights from neuroscience and psychology to try to make Facebook feel like a safer, more benevolent place for adults and kids alike—and even help users resolve conflicts online that they haven’t been able to tackle offline.
“Essentially, the problem is that Facebook, just like any other social context in everyday life, is a place where people can have conflict,” says Paul Piff, a postdoctoral psychology researcher at UC Berkeley who is working on the project, “and we want to build tools to enable people who use Facebook to interact with each other in a kinder, more compassionate way.”
Facebook as relationship counselor
For users troubled by a photo, Facebook provides the option to click a Report link, which takes them through a sequence of screens where they can elaborate on the problem, called the “reporting flow.”
Up until a few months ago, the flow presented all “reporters” with the same options for resolving the problem, regardless of what that problem actually was; those resolutions included unfriending the user or blocking him or her from ever making contact again on Facebook.
“One thing that we learned is that if you give someone the tool to block, that’s actually not in many cases the right solution because that ends the conversation and doesn’t necessarily resolve anything—you just sort of turn a blind eye to it,” says Jacob Brill, a product manager on Facebook’s Site Integrity and Support Engineering team, which tries to fix problems users are experiencing on the site, from account fraud to offensive content.

Instead, Brill’s team concluded that a better option would be to facilitate conversations between a person reporting content and the user who uploaded the content, a system that they call “social reporting.”
“I really think that was key—that the best way to resolve conflict on Facebook is not to have Facebook step in, but to give people tools to actually problem-solve themselves,” says Piff. “It’s like going to a relationship counselor to resolve relationship conflict: Relationship counselors are there to give couples tools to resolve conflict with each other.”
To help Facebook develop those tools, Bejar turned to Piff and two of his UC Berkeley colleagues, social psychologist Dacher Keltner and neuroscientist Emiliana Simon-Thomas—the GGSC’s faculty director and science director, respectively—all of whom are experts in the psychology of emotion.
“It felt like we could sharpen their communication,” says Keltner, “just to make it smarter emotionally, giving kids and adults sharper language to report on the complexities of what they were feeling.”
The old reporting flow wasn’t very emotionally intelligent. When first identifying the problem to Facebook, users had some basic options: They could select “I don’t like this photo of me,” claim that the photo was harassing them or a friend, or say that it violated one of the site’s Community Standards—for hate speech or drug use or violence or some other offense. Then they could unfriend or block the other user, or send that user a message.
Initially, users had to craft that message themselves, and only 20 percent of them actually sent a message. To boost that rate, Facebook provided some generic default text—“Hey I don’t like this photo. Please remove it.”—which raised the send rate to 51 percent. But often users would send one of these messages and never hear back, and the photo wouldn’t get deleted.
Bejar, Brill, and others at Facebook thought they could do better. The Berkeley research team believed this flow was missing an important step: the opportunity for users to identify and convey their emotions. That would guard against the fact that it’s easier for people online to be insensitive or even oblivious to how their actions affect others.
“If you get someone to express more productively how they’re feeling, that’s going to allow someone else to better understand those feelings, and try to address their needs,” says Piff. “There are some very simple things we can do to give rise to more productive interpersonal interactions.”

Instead of simply having users click “I don’t like this photo,” for instance, they decided to prompt users with the sentence “I don’t like this photo because:”, which they could complete with emotion-laden phrases, such as “It’s embarrassing” or “It makes me sad” (see screenshot at left). People reporting photos selected one of these options 78 percent of the time, suggesting that the list of phrases effectively captured what they were feeling.
People were then taken to a screen telling them that the best way to remove the photo was to ask the other user to take it down—blocking or unfriending were no longer presented as options—and they were given more emotionally intelligent text for a message they could send through Facebook, tailored to the particular situation.

That text included the other person’s name, asked him or her more politely to remove the content (“would you please take it down?” vs. the old “please remove it”), and specified why the user didn’t like the photo, emphasizing their emotional reaction and point of view—but still keeping a light touch. For example, photos that made someone embarrassed are described as “a little embarrassing to me.” (See the screenshot at left for an example.)
It worked. Roughly 75 to 80 percent of people in the new, emotionally intelligent flow sent these default messages without revising or editing the text, a 50 percent increase from the number who sent the old, impersonal message.
When Keltner and his team presented these findings at Facebook’s second Compassion Research Day, a public event held on Facebook’s campus earlier this month, he emphasized that what mattered wasn’t just that more users were sending messages but that they were enjoying a more positive overall experience.
“There are a lot of data that show when I feel stressed out, mortified, or embarrassed by something happening on Facebook, that activates old parts of the brain, like the amygdala,” Keltner told the crowd. “And the minute I put that into words, in precise terms, the prefrontal cortex takes over and quiets the stress-related physiology.”
Preliminary data seem to back this up. Among the users who sent a message through this new flow, roughly half said they felt positively about the other person (called the “content creator”) after they sent him or her the message; less than 20 percent said they felt negatively. (The team is still collecting and analyzing data on how users feel before they send the messages, and on how positively they feel after sending a message through the old flow.)
In this new social reporting system, half of content creators deleted the offending photo after they received the request to remove it, whereas only a third deleted the photo under the old system. Perhaps more importantly, roughly 75 percent of the content creators replied to the messages they received, using new default text that the researchers crafted for them. That’s a nearly 50 percent increase from the number who replied to the old kinds of messages.
“The right resolution isn’t necessarily for the photo to be taken down if in fact it’s really important to the person who uploaded it,” says Brill. “What’s really important is that you guys are talking about that, and that there is a dialogue going back and forth.”
This post is a problem
That’s all well and good for Facebook’s adult users, but kids on Facebook often need more. For them, Facebook’s hazards include cyberbullying from peers and manipulation by adult predators. Rough estimates indicate that more than half of kids have had someone say mean or hurtful things to them online.
Previously, if kids felt hurt or threatened by someone on Facebook, they could click the same Report link adults saw, which took them through a similar flow, asking if they or friends were being “harassed.” From there, Facebook gave them the option to block or unfriend that person and send him or her a message, while also suggesting that they contact an adult who could help.

But after hearing Yale developmental psychologist Marc Brackett speak at the first Compassion Research Day in December of 2011, Bejar and his colleagues realized that the old flows failed to acknowledge the particular emotions that these kids were experiencing. That oversight might have made the kids less likely to engage in the reporting process and contact a supportive adult for guidance.
“The way you really address this,” Bejar said at the second Compassion Research Day, “is not by taking a piece of content away and slapping somebody’s hand, but by creating an environment in which children feel supported.”
To do that, he enlisted Brackett and two of his colleagues, Robin Stern and Andres Richner. The research team organized focus groups with 13-to-14-year-old kids, the youngest age officially allowed on Facebook, and interviewed kids who’d experienced cyberbullying. The team wanted to create tools that were developmentally appropriate to different age ranges, and they decided to target this youngest group first, then work their way up.
From talking with these adolescents, they pinpointed some of the problems with the language Facebook was using. For instance, says Brackett, some kids thought that clicking “Report” meant that the police would be called, and many didn’t feel that “harassed” accurately described what they had been experiencing.
Instead, Brackett and his team replaced “Report” with language that felt more informal: “This post is a problem.”
They tried to apply similar changes across the board, refining language to make it more age-appropriate. Instead of simply asking kids whether they felt harassed, they enabled kids to choose among far more nuanced reasons for reporting content, including that someone “said mean things to me or about me” or “threatened to hurt me” or “posted something that I just don’t like.” They also asked kids to identify how the content made them feel, selecting from a list of options.
Depending on the problem they identified, the new flows gave kids more customized options for the action they could take in response. That included messages they could send to the other person, or to a trusted adult, that featured more emotionally rich and specific language, tailored to the type of situation they were reporting.

“We wanted to make sure that they didn’t feel isolated and alone—that they would receive support in a way that would help them reach out to adults who could provide them with the help that they needed,” Brackett said when presenting his team’s work at the second Compassion Research Day.
After testing these new flows over two months, the team made some noteworthy discoveries. One surprise was that, when kids reported problems that they were experiencing themselves, 53 percent of those problems concerned posts that they “didn’t like,” whereas only three percent of the posts were seen as threatening.
“The big takeaway here is that … a lot of the cases are interpersonal conflicts that are really best resolved either between people or with a trusted adult just giving you a couple of pointers,” Jacob Brill said at the recent Compassion Research Day. “So we’re giving [kids] the language and the resources to help with a situation.”
And those resources do seem to be working: Forty-three percent of kids who used these new flows reached out to a trusted adult when reporting a problem, whereas only 19 percent did so with the old flows.
“The new experience that we’re providing is empowering kids to reach out to someone they trust to get the help that they need,” says Brackett. “There’s nothing more gratifying than being able to help the most amount of kids in the quickest way possible.”
Social reporting 2.0
Everyone involved in the project stresses that it’s still in its very early stages. So far, it has only targeted English-language Facebook users in the United States. Brackett’s team’s work has only focused on 13 to 14 year olds, and the new flows developed by the Berkeley team were only piloted on 50 percent of Facebook users, randomly selected.
Can they build a more emotionally intelligent form of social reporting that works for different cultures and age groups?
“Our mission at Facebook is to do just that,” says Brill. “We will continue to figure out how to make this work for anyone who has experiences on Facebook.”
The teams are already working to improve upon the results they presented at the second Compassion Research Day. Brackett says he believes they can encourage even more kids on Facebook to reach out to trusted adults, and he’s eager to start helping older age groups. And he’s excited by the potential impact of this work.
“When we do our work in schools, it’s one district, one school, one classroom at a time,” he says. “Here, we have the opportunity to reach tens of thousands of kids.”
And that reach carries exciting scientific implications for the researchers.
“We’re going to be the ones who get to go in and have 500,000 data points,” says Simon-Thomas. “It’s beyond imagination for a research lab to get that kind of data, and it really taps into the questions we’re interested in: How does conveying your emotion influence social dynamics in rich and interesting ways? Does it facilitate cooperation and understanding?”
And what’s in it for Facebook?
Bejar, the father of two young children, says that protecting kids and strengthening connections between Facebook users makes the site more self-sustaining in the long run. The project will have succeeded, he says, if it encourages more users to think twice before posting a photo that might embarrass a friend, or even to notify that friend when they post a questionable image.
“It’s those kinds of kind, compassionate interactions,” he says, “that help build a sustainable community.”
Source: & Facebook Community.