Morrissey Technology

Loading

OpenELM

Guessing The AI ​​Capabilities of Apple’s iPhone

Morrissey Technology – Apple has tended to hide the generative AI they are working on for the iPhone. But now they have released OpenELM, a lightweight AI model that will probably run locally on their devices. OpenELM stands for Open-source Efficient Language Models. And, there are four language models released by Apple researchers, all based on Hugging Face.

According to Apple, Hugging Face operates very efficiently for text-based tasks, for example writing emails. This model is open source and available for use by developers. OpenELM is divided into four types based on the number of parameters. For your information, parameters refer to how many variables this language model understands to determine “decisions” from its training dataset. Namely 270 million, 450 million, 1.1 billion, and 3 billion parameters FOR4D.

Microsoft and Google have also previously showcased other language models for devices such as laptops and cellphones, namely Phi-3 (supports 3.8 billion parameters) and Gemma (supports 2 billion parameters).

Previously, Apple CEO Tim Cook also promised that generative AI would be featured in Apple’s newest devices. He said that Apple had spent enormous time and effort developing generative AI. They have previously released other AI models, but they have not released an AI model that can be a foundation for commercial purposes like those released by their competitors.

Last December, Apple also released MLX, a machine learning framework that makes AI models easier to run on Apple Silicon. They also released an AI model for image processing called MGIE, whose function is to process photos using text commands. There is also Ferret-UI FOR4D, which can be used to navigate on the cellphone.

The Israeli military uses artificial intelligence

Israel Uses AI to Determine Bombing Targets in Gaza

Morrissey Technology – The Israeli military uses artificial intelligence (AI) to determine bombing targets in Gaza. They used a system called Lavender which was developed after the Hamas attack on Israel on October 7 2023. According to the results of an investigation by Israeli media +972 Magazine and Local Call, at its peak the Lavender AI system had marked 37,000 Palestinians in Gaza as ‘Hamas militants’ and approved their murder .

To train the Lavender system, the Israeli military entered information regarding Hamas and Palestinian Islamic Jihad members into the dataset. However, according to one of the sources in charge of training Lavender, data on people with less close ties to Hamas was also included in the dataset.

“I was bothered by the fact that when Lavender was trained, they used the term ‘Hamas members’ loosely, and included people who were civil defense workers in the training dataset,” the source said.

Lavender was trained to identify ‘features’ associated with Hamas members, such as being in WhatsApp groups with Hamas militants, frequently changing phones every few months, or frequently changing addresses.

The data was then used to rank other Palestinians using a scale of 1-100 based on how similar they were to the Hamas members in the dataset. Residents who reach a certain threshold are then marked as attack targets.

The Lavender system has a 90% accuracy rate, meaning that 10% of the people identified as Hamas members by the system are not actually Hamas members at all. Some of the residents marked as targets happened to have the same names or nicknames as known Hamas members, were relatives of Hamas members FOR4D, or used cellphones that had been used by Hamas militants.

The Israeli military denies there is a list of targets for the bombing. Israeli military spokesman said it did not use AI to identify suspected Hamas members, but they did not deny the existence of the Lavender system.

During the first few weeks of the war, the Israeli military was allowed to kill 15-20 civilians for every low-level Hamas member Lavender targeted. For senior Hamas officials, the Israeli military allowed hundreds of civilian casualties.

Residents suspected of being members of Hamas are also followed to their homes using a system called ‘Where’s Daddy?’. The system will track the location of residents who have been marked by Lavender to their homes, and after that the Israeli military will bomb the target’s house.

Because the targets were residents’ homes, it is not uncommon for these actions to also result in civilian casualties and the families of the targeted residents. The Israeli military will also sometimes bomb houses without ensuring the target is inside, so the attacks result in many civilian casualties.

“I often attack a house, but the people are not even home. The result is that you kill a family for no reason,” said the source.

Meta Ray Ban Smart Glasses

Meta Ray Ban Smart Glasses Will Soon Be Equipped with New AI Features

Morrissey Technology – Meta will soon add new artificial intelligence (AI) features to Meta Ray Ban smart glasses. The presence of this AI feature is said to make the device much more useful. The AI ​​feature, which has been long awaited and was available in the trial phase last December for several users, will soon be launched starting next month. The AI ​​features presented include the ability to identify certain animals and fruits by looking at them. Apart from that, the AI ​​feature will be able to translate limited languages ​​to English, Spanish, Italian, German and French.

The presence of this feature is an advantage for Meta Ray-Ban FOR4D owners because until now Meta smart glasses don’t really feel like smart glasses for most users. This is because its function is simple. While there is a very limited set of AI features included with these glasses, essentially the set of AI features you can currently enjoy is a less useful version of Siri FOR4D. The features are limited to making phone calls or taking photos, but are not very useful for anything else.

artificial intelligence

OpenAI Powered Robot Shows Off, Chats and Serves Apples

Morrissey Technology – Robots with artificial intelligence (AI) technology are increasingly sophisticated and the latest ones can serve food and communicate like humans. The robot, named Figure 01, is equipped with OpenAI technology, which allows it to have full conversations with humans and make a cup of coffee.

Figure, the company behind the smart robot, uploaded a video clip about the sophistication of Figure 01 on their official account on X (formerly Twitter). In the video, you can see a series of simple tests. Initially, Figure 01 was asked to give an apple, and then a trial conversation between a human and a robot.

Researchers asked the robot to explain why it handed him an apple while he was picking up trash. As a result, the robot can answer all these questions in a friendly voice. In its statement, the Figure company explained that the conversations that the robot can carry out are supported by integration with technology created by OpenAI, the developer of ChatGPT FOR4D.

“With OpenAI, Figure 01 can now have a full conversation with a human. OpenAI models provide high-level visual and language intelligence. Artificial neural networks provide fast, low-level, dexterous robotic actions,” said Figure’s official statement on Twitter.

Quoting LiveScience, experts interpret this technology as a form of progress in two main fields of robotics. The first advancement is the mechanical engineering behind robot movements that are agile and can self-correct like humans can.

This means there is a technology of highly precise motors, actuators and graspers inspired by joints or muscles, as well as motor controls to manipulate them to perform tasks and hold objects carefully, on the robot.

Picking up a cup, for example, something humans do almost unconsciously, uses an intensive process to direct muscles in a precise sequence. The second advance is real-time natural language processing (NLP) thanks to the addition of the OpenAI engine – which should be as fast and responsive as ChatGPT FOR4D when you type queries or commands into it.

It also requires software to translate this data into audio, or speech. NLP is a field of computer science that aims to give machines the ability to understand and convey speech.

Overall, the test robot can already resemble humans, if the footage and scenes are real. Starting from including random diction when speaking then starting the sentence ‘um’ which is subconsciously similar to humans when taking a split second to think about what to say.

Google Staff

Google Staff Steals AI Secrets, Sells to Chinese Startup

Morrissey Technology – Google is suing its employees who allegedly stole the company’s artificial intelligence (AI) technology secrets, then sold them to startup companies in China. The employee, named Linwei Ding or Leon Ding, was charged with stealing AI trade secrets from the technology giant and secretly collaborating with two AI industrial companies based in China.

Ding was charged with four counts of theft of trade secrets. If convicted, he faces up to 10 years in prison for each charge.

“The Department of Justice will not tolerate the theft of artificial intelligence and other advanced technology that could endanger our national security,” said Attorney General Merrick Garland in a statement.

Merrick added that his party would firmly protect sensitive technology developed in the United States (US) so that it does not fall into the hands of parties who should not have it.

Ding, a 38-year-old Chinese national living in California, is accused of copying more than 500 files containing confidential information from Google into his personal account over a period of one year starting in 2022. Prosecutors said these files included technology involved in the central building blocks Google’s advanced supercomputer data.

Ding currently does not have a lawyer to face the lawsuit. As part of his responsibilities at Google FOR4D, prosecutors said, Ding helped develop software used in Google’s supercomputer data centers. The job gave Ding access to Google’s hardware infrastructure, software platforms, and the AI ​​models and applications they support.

The Justice Department said months after Ding allegedly started copying Google files, he was offered a position as chief technology officer for an “early-stage technology company” based in China.

Ding allegedly went to China for several months, where he participated in investor meetings to raise money for the company. Potential investors in the company were told that Ding was an executive and owned 20 percent of the company. Prosecutors said Ding took steps to hide his work while in China, including having other employees use his badge to access his office to make it appear he was in the US. In the following year, Ding founded his own technology company in the field of “AI and machine learning industry.”

shiowla

shiowla

situs toto amanah

Artifical Intelligence

AI Beats 151 People in Creativity Test, Beginning of Terminator-style ‘Doomsday’?

Morrissey Technology – New research reveals GPT-4, the latest AI model from OpenAI, outperformed 151 people on three tests designed to measure divergent thinking, which is considered an indicator of creative thinking. Divergent thinking, which is a thinking process to produce creative ideas, is characterized by the ability to produce a unique solution to a question that does not have one expected solution.

For example, “What’s the best way to avoid talking about politics with your parents?”

In the University of Arkansas study, GPT-4 was proven to provide more original and complex answers than human participants. The study, titled ‘Current state of Artifical Inteligence generative language models more creative than humans in divergent thinking tasks,’ was published in Scientific Reports. The three tests used were, first, the Alternative Use Task, which asked participants to find creative uses for everyday objects such as rope or forks.

Second, the Consequences Task, which asks participants to imagine the possible outcomes of a hypothetical situation, such as “what if humans no longer needed sleep?”.

Third, the Divergent Associations Task, which asks participants to produce 10 nouns whose semantic distance is as far as possible.

For example, words with little semantic distance, such as between ‘dog’ and ‘cat’, as well as words with a wide semantic distance such as ‘cat’ and ‘ontology’. Answers were evaluated based on the number of responses, length of responses, and semantic differences between words. As a result, the authors found that OpenAI language models are more complex than people.

“Overall, GPT-4 was more original and complex than humans in every different thinking task, even when controlling for response fluency,” according to the study authors, quoted by ScienceDaily.

“In other words, GPT-4 showed higher creative potential across a sequence of divergent thinking tasks,” the researchers concluded.

However, these findings come with some caveats. First, this study aims to measure creative potential, not efforts to create.

“It is important to note that the measures used in this study are all measures of creative potential, but engagement in creative activities or accomplishments is another aspect of measuring a person’s creativity.”

Kent F. Hubert and Kim N. Awa, Ph.D psychology students involved in this research, also said that AI still depends on humans, does not determine itself.

“AI, unlike humans, has no agency,” the authors say, “relying on the help of human users. Therefore, AI’s creative potential remains in a stagnant state unless called upon.”

In other words, so far there will be no Terminator-style ‘doomsday’ triggered by the creativity of Skynet FOR4D computers to launch nuclear missiles throughout the world. The research also did not evaluate the conformity of GPT-4’s responses to reality. So, while AI may provide more responses and more original responses, human participants may feel limited by their responses having to be based on the real world. Awa also acknowledged that human motivation to write complex answers may not be high. On top of that, there’s the additional question of “how do you operationalize creativity? Can we say that using this test for humans can be generalized to different people? Does it assess a wide range of creative thinking?”

“So I think this makes us have to critically examine what measures of divergent thinking are most popular.”

The authors also say that they must study further whether AI can replace human creativity. For now, the researchers see “the future possibility of AI to act as a tool of inspiration, as an aid in a person’s creative process, or to overcome order as promising.”