It is important because police departments are starting to use drones for a lot of things. San Francisco approved the use of drones that can kill people in emergencies last week. Most police drones in the UK have thermal cameras that can be used to detect how many people are in a house. This has been used to catch human traffickers, as well as to target people holding suspected parties during covid-19 lock downs.

Psch says that virtual reality will allow researchers to test the technology in a safe way.

I found the encounter with the drones unnerving even though I knew I was in a virtual reality environment. Even though I met a human-operated one, my opinion of these drones did not change.

Christian Enemark, a professor at the University ofSouthampton who specializes in the ethics of war and drones and is not involved in the research, says that it may not make a difference whether drones areolite or rude. The use of drones is a reminder that the police aren't here, whether they're not bothering to be here or they're too afraid to be here.

Maybe there is something disrespectful about an encounter.

Deeper Learning

GPT-4 is on the way, but Openai is still fixing GPT-3.

The internet is abuzz with excitement about the latest iteration of Openai's large language model, GPT 3. The demo answers people's questions in a back and forth dialogue. Over one million users have used the demo since it was launched last week. Here you can read Will Douglas Heaven's story.

GPT3 is a confident bull and can easily be asked to say toxic things. Openai says that it has fixed a lot of the problems with its software, which answers follow-up questions, admits its mistakes, and rejects inappropriate requests. It won't answer questions such as how to be evil or how to break into someone's house